Introducing Qwen2: Next-Generation AI Models
Introducing Qwen2: Next-Generation AI Models
After extensive development, we proudly present Qwen2, an upgraded suite from Qwen1.5. This release includes models like Qwen2-0.5B to Qwen2-72B, boasting enhanced multilingual capabilities, faster performance, and an extended context length up to 128K tokens. These models are designed to excel in coding, mathematics, and various benchmark evaluations, ensuring top-notch performance.
Qwen2 models have been pre-trained and instruction-tuned across a diverse linguistic dataset, incorporating 27 languages beyond English and Chinese. This diverse training enhances their proficiency in multilingual contexts and code-switching scenarios, providing robust and reliable outputs across different languages and tasks.
Significant improvements are evident in Qwen2-72B's performance, outperforming leading models such as Llama-3-70B. Post-training enhancements focus on areas like coding, mathematics, reasoning, and multilingual understanding, aligning the models with human values for safe and effective use. Innovative training techniques and scalable methods ensure high-quality, diverse, and creative outputs with minimal human annotation.
With the open-sourcing of these models on platforms like Hugging Face and ModelScope, the Qwen2 series is accessible for various applications. Future developments include larger models and multimodal capabilities, emphasizing the commitment to advancing AI research and practical implementations within the community.