RLHF
Reinforcement Learning from Human Feedback (RLHF) is a technique that uses human feedback to train reinforcement learning (RL) agents.
Read MoreSelf-Play Fine-tuning (SPIN) is a new fine-tuning method for Large Language Models (LLMs) that can significantly improve performance without the need for additional human-annotated data.
Read More