Chapter 6. Advanced Fine-tuning Techniques
In the previous chapter, we presented the canonical way to fine-tune a typical LLM. In the real world, there are a wide variety of motivations for updating an LLM, and similarly there are multiple ways to update it. In this chapter, we will describe several advanced fine-tuning techniques, and highlight the scenarios in which each technique would be suitable.
Why would you want to update the parameters of an LLM? We touched upon this in previous chapters but let’s go through it in more detail now:
-
Domain Adaptation: The data that we work with belongs to a specialized domain that the LLM might not have been familiarized with during pre-training. In this case, we would like to update the model by training it on domain-specific data.
-
Task Adaptation: We care about LLM performance on specific downstream tasks. In order to improve the LLM’s performance on these tasks, we can train it on ...
Get Designing Large Language Model Applications now with the O’Reilly learning platform.
O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.