LLMJuly 15, 2024Alibaba Cloud / Qwen Team
Qwen2 Technical Report
Alibaba Cloud
Abstract
We introduce Qwen2, the next generation of the Qwen series of large language models. Qwen2 includes dense language models of 0.5B, 1.5B, 7B, 57B-A14B (MoE), and 72B parameters, trained on data in 29 languages. Qwen2-72B achieves competitive performance with leading proprietary models on a wide range of benchmarks.
Key Findings
- 1Released a family of models from 0.5B to 72B parameters
- 2Trained on data in 29 languages for strong multilingual support
- 3Included a 57B MoE model with only 14B active parameters
- 4Achieved performance competitive with leading proprietary models
- 5Released as open source with commercial-friendly licensing
Impact & Significance
Qwen2 established Alibaba as a leading contributor to open-source AI, providing high-quality multilingual models that are widely adopted in Asia and globally through platforms like Ollama and Hugging Face.
Related Tools
Related Papers
LLMJuly 23, 2024
The Llama 3 Herd of Models
Meta AI
EfficiencyMay 7, 2024
DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model
DeepSeek AI
LLMMarch 4, 2024
The Claude 3 Model Family: Opus, Sonnet, and Haiku
Anthropic
VisionFebruary 15, 2024
Video Generation Models as World Simulators
OpenAI