Glossary

RLHF

Reinforcement Learning from Human Feedback (RLHF) is a technique that uses human feedback to train reinforcement learning (RL) agents.

Read More

LoRAMoE

LoRAMoE is a plugin version of Mixture of Experts (MoE) that can effectively prevent world knowledge forgetting in large language models (LLMs) during supervised fine-tuning.

Read More

Self-Play Fine-tuning (SPIN)

Self-Play Fine-tuning (SPIN) is a new fine-tuning method for Large Language Models (LLMs) that can significantly improve performance without the need for additional human-annotated data.

Read More

DPO

Direct Preference Optimization (DPO) is a simplified and efficient approach to fine-tuning large language models (LLMs).

Read More

Arc-c

ARC-c is a challenging variation of the ARC Benchmark, designed to assess the reasoning and commonsense understanding of large language models. Learn more about this dataset and the difficulty it presents for models.

Read More

Arc-e

ARC-e is an enhanced version of the ARC Benchmark, evaluating large language models’ reasoning abilities. With 1,169 challenging questions, no model has reached a 75% score yet.

Read More