Yi-1.5-34B

by | May 13, 2024

Yi-1.5 is a state-of-the-art AI model, designed to revolutionize the way we interact with technology. It’s been trained on a massive 500 billion token corpus, and fine-tuned with 3 million diverse samples. This extensive training allows Yi-1.5 to excel in coding, math, reasoning, and instruction-following tasks. The model’s enhanced capabilities make it an invaluable tool for a wide range of applications, from software development to academic research. Experience the future of AI with Yi-1.5.

Yi-1.5 is a groundbreaking AI model, pushing the boundaries of what’s possible in the realm of artificial intelligence. It’s been trained on an enormous corpus of 500 billion tokens, encompassing a wide array of fields and disciplines. This extensive pre-training has equipped Yi-1.5 with a comprehensive understanding of language, context, and knowledge.

Beyond language comprehension, Yi-1.5 excels in tasks involving coding, math, reasoning, and instruction-following, thanks to fine-tuning on 3 million diverse samples. Whether you’re a software developer automating coding tasks, a researcher analyzing complex datasets, or an educator creating interactive learning experiences, Yi-1.5 is ready to assist.

What sets Yi-1.5 apart is its adaptability. The model can be further fine-tuned for specific tasks or domains, making it a versatile tool for a broad range of applications. With Yi-1.5, the future of AI is here and now, ready to be explored and utilized.

Current
Apache License
Pretrained, Instruction-tuned

Comparison 

Sourced on: May 13, 2024

General Performance

  • MMLU: Yi-1.5-34B-Chat shows a score of 76.8. Other models like Qwen1.5-32B-Chat, Qwen1.5-72B-Chat, Mixtra1-Bx78-Instruct-v0.1, and Mixtra1-Bx228-Instruct-v0.1 scored 74.3, 77.3, 71.4, and 77.7 respectively.
  • GSM8K: Yi-1.5-34B-Chat scores 90.2, outperforming all other models.
  • MATH: Yi-1.5-34B-Chat scores 50.1, leading the pack with the highest score.
  • HumanEval: Yi-1.5-34B-Chat achieves 75.2, only surpassed by Mixtra1-Bx228-Instruct-v0.1 with 76.2.
  • MBPP: Yi-1.5-34B-Chat reaches 74.6, again only surpassed by Mixtra1-Bx228-Instruct-v0.1 with 73.8.
  • MT-Bench: Yi-1.5-34B-Chat and Qwen1.5-72B-Chat both score 8.5, the highest in this category.
  • Alignbench: Yi-1.5-34B-Chat achieves 7.2, tied with Qwen1.5-72B-Chat for the highest score.
  • ArenaHard: Yi-1.5-34B-Chat scores 42.6, the highest in this category.
  • AlpacaEval2.0: Yi-1.5-34B-Chat and Qwen1.5-72B-Chat both score 36.6, the highest in this category.

Highlights:

  • Yi-1.5-34B-Chat consistently performs well across all categories, often scoring the highest or near the top.
  • The most significant lead is observed in the GSM8K category, where Yi-1.5-34B-Chat outperforms all other models.
  • Yi-1.5-34B-Chat shows strong performance in knowledge reasoning and reading comprehension tasks, indicating a substantial advancement over other models.
BenchmarkYi-1.5-34B-ChatQwen1.5-32B-ChatQwen1.5-72B-ChatMixtral-8x7B-Instruct-v0.1Mixtral-8x22B-Instruct-v0.1
MMLU76.874.377.371.477.7
GSM8K90.283.986.065.784
MATH50.143.344.430.741.1
HumanEval75.251.264.645.176.2
MBPP74.666.972.559.573.8
MT-Bench8.58.38.68.38.6
Alignbench7.27.17.25.76.5
ArenaHard42.624.236.123.436.4
AlpacaEval2.036.623.936.623.730.9

Team 

The team behind the LLM model is a diverse group of highly skilled individuals, each bringing their unique expertise to the table. The team comprises data scientists, machine learning engineers, and researchers who have extensive experience in the field of artificial intelligence. They have backgrounds in various disciplines such as computer science, mathematics, and linguistics, which allows them to approach the development of the LLM model from multiple perspectives. The team’s collaborative efforts have resulted in the creation of a model that is capable of understanding and generating human-like text, making significant strides in the field of natural language processing. Their dedication to continuous learning and improvement ensures that the LLM model remains at the forefront of AI technology. The team’s work is guided by a commitment to ethical AI practices, ensuring that the LLM model is used responsibly and benefits a wide range of users. Their collective expertise and dedication have been instrumental in the successful development and deployment of the LLM model. They continue to innovate and push the boundaries of what is possible in AI technology.

Community 

Community engagement through various channels such as GitHub, Discord, WeChat, and the Yi Learning Hub. It also invites users to contribute to the model’s development and participate in discussions, reflecting a strong emphasis on collaborative improvement and open-source ethos.

Active Members: 51-100 Members
Engagement Level: High Engagement