Qwen2: Advanced Multilingual and Computational Abilities

by | Jun 23, 2024

Qwen2 is an advanced AI model series with sizes ranging from 0.5B to 72B parameters, trained on data in 29 languages. It offers state-of-the-art performance in benchmarks, excels in coding and math, and supports up to 128K token contexts. Models are open-sourced for community development, with a focus on safety and multilingual capabilities.

Introducing Qwen2, the latest advancement in AI technology, designed to cater to a wide range of computational needs. Qwen2 is a series of pretrained and instruction-tuned models, available in five sizes: Qwen2-0.5B, Qwen2-1.5B, Qwen2-7B, Qwen2-57B-A14B, and Qwen2-72B. These models have been meticulously trained on data in 29 languages, ensuring state-of-the-art performance across various benchmarks.

Qwen2 models boast significant improvements in coding and mathematics, with extended context length support up to 128K tokens for certain models. The use of Group Query Attention (GQA) across all sizes enhances speed and reduces memory usage during inference3. For smaller models, tie embedding is employed to manage the large sparse embeddings efficiently.

The Qwen2 series demonstrates exceptional multilingual capabilities, having been trained on 27 additional languages besides English and Chinese. This training includes a focus on code-switching scenarios, which are common in multilingual communication, resulting in improved model performance.

Safety and responsibility are paramount in the design of Qwen2. The models have been evaluated against harmful responses in multiple languages, showing comparable safety performance to other leading models. Qwen2’s licensing has also evolved, with most models adopting the Apache 2.0 license, promoting openness and accelerating global applications.

Qwen2 is not just a technological leap forward; it’s a commitment to responsible AI development, offering enhanced performance, safety, and accessibility to the AI community. With Qwen2, users can expect a reliable, efficient, and ethically aligned AI experience.

Current
Tongyi Qianwen LICENSE AGREEMENT
Instruction-tuned

Comparison 

Sourced on: June 7, 2024

Overall Performance: Qwen2-72B shows superior performance across most benchmarks compared to Llama3-70B, Mixtral-8x22B, and Qwen1.5-110B.

  • Natural Language Understanding: Excels in MMLU with 84.2%, outperforming others.
  • Advanced Reasoning: Leads in TheoremQA at 43.1%.
  • Coding Proficiency: Dominates in HumanEval with 64.6%.
  • Mathematical Skills: Tops in GSM8K at 89.5%.
  • Multilingual Capabilities: Achieves 91% in C-Eval and 90.1% in CMMLU, indicating strong performance in Chinese language tasks.

Qwen2-72B demonstrates a significant advancement in language model capabilities, particularly in coding and mathematics, while also maintaining robust multilingual abilities1.

BenchmarkQwen2-72BLlama3-70BMixtral-8x22BQwen1.5-110B
MMLU84.279.577.880.4
MMLU-Pro55.652.849.549.4
GPQA37.936.334.335.9
TheoremQA43.132.335.934.9
BBH82.48178.974.8
HumanEval64.648.246.354.3
MBPP76.970.471.770.9
MultiPL-E59.646.346.752.7
G5M8K89.58383.785.4
MATH51.142.541.749.6
C-Eval9165.254.689.1
CMMLU90.167.253.488.3
Multi-Exam76.67063.575.6
Multi-Understanding80.779.977.778.2
Multi-Mathematics7667.162.964.4

Team 

The Qwen team is dedicated to advancing artificial general intelligence, focusing on the development of generalist models such as large language models and multimodal models. They support open-source initiatives and have released a series of models including Qwen-7B, Qwen-14B, Qwen-72B, and their chat and multimodal counterparts, Qwen-VL and Qwen-Audio. Additionally, they offer web services and an app to assist users in their daily work and life. The team comprises individuals with diverse talents and interests and encourages engagement and new members.

https://qwenlm.github.io/publication/

Community 

The Qwen community actively contributes to the development and improvement of Qwen models on GitHub. The official repository, maintained by the Qwen Team, hosts pretrained large language models proposed by Alibaba Cloud. While the Qwen/Qwen repo is no longer actively maintained due to substantial codebase differences, the community continues to engage with Qwen2 and share experiences. Qwen2, along with its chat variants, offers strong language capabilities, multilingual support, and impressive performance across various tasks. The team encourages collaboration and feedback from the community, making Qwen a dynamic and evolving project.

Active Members: 5001-10,000 Members
Engagement Level: High Engagement

Resources

List of resources related to this product.