Introduction to Fine-Tuning

LLMs are pre-trained on huge amounts of text data to learn general language patterns.

LLMs can be fine-tuned on a much smaller amount of data to excel at a particular task (e.g., classification of financial text).

Note: LLM pre-training is generally unsupervised.

Types of Fine-Tuning

Fine-tuning can be a supervised or unsupervised process and involves:

  • Changing some of the LLM weights,
  • Changing all of the LLM weights (full fine-tuning), or
  • Parameter Efficient Fine-Tuning (PEFT), i.e., keeping the LLM weights the same but updating a small number of additional parameters that will adjust the LLM weights (e.g., LoRA).

The more weights you update, the more computational resources you need.

Previous
Next
RC Logo © 2025 The Rector and Visitors of the University of Virginia