A LoRA-Based Approach to Fine-Tuning LLMs for Educational Guidance in Resource-Constrained Settings

Read the full article See related articles

Discuss this preprint

Start a discussion What are Sciety discussions?

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

The current study describes a cost-effective method for adapting large language models(LLMs) for academic advising in study-abroad contexts. Using the Mistral-7B model withLow-Rank Adaptation (LoRA) and 4-bit NF4 quantization via the Unsloth framework, themodel underwent training in two distinct hardware phases to demonstrate adaptability andcomputational efficiency. Contrary to multi-stage data curation, this study utilized a sin-gle synthetic dataset of 2,274 conversation pairs across both phases to evaluate hardware-agnostic convergence. In Phase 1, the model was fine-tuned on an NVIDIA Tesla P100.In Phase 2, the model continued training on the same dataset using an NVIDIA Tesla T4with optimized batch configurations to refine performance. Technical innovations utilizedmemory-efficient quantization and continuous training analytics. After training, the studydemonstrated a total reduction in training loss from ∼1.01 to ∼0.34, achieving stable conver-gence on consumer-grade GPU equipment. These findings support the effective applicationof instruction-tuned LLMs within educational advising, specifically showing that trainingcan be effectively distributed and resumed across varying hardware architectures.

Article activity feed