Supervised fine-tuning (SFT) is a frequently used method for aligning large language models (LLMs) to human preferences. It involves curating a dataset of high-quality LLM outputs and then fine-tuning the model on this data using a next token prediction objective.
Supervised fine-tuning (SFT) is a powerful tool for aligning LLMs to human preferences and making them more useful and reliable.