March 2026
Intermediate to advanced
402 pages
11h 1m
English
In the previous chapter, we introduced the Transformer model's encoder-decoder architecture and how to use the Transformers library and pretrained models to fine-tune on specific datasets. Here, we will further discuss the approach of parameter-efficient fine-tuning (PEFT), introduce the Transformer Reinforcement Learning (TRL) library for explaining the supervised fine-tuning (SFT) process, and discuss in detail various techniques such as low-rank adaptation (LoRA) for fine-tuning, evaluation of PEFT models, and the workflow.
Fine-tuning a large pretrained model involves updating the weights of the model to refine its policy for performing specific tasks. The number of parameters to update is high, making the ...
Read now
Unlock full access