Qwen 2.5

Qwen 2.5

In this blog, we explore the details of the new Qwen2.5 series language models developed by the Alibaba Cloud Dev Team. The team has created a range of decoder-only dense models, with seven of them being open-sourced, ranging from 0.5B to 72B parameters. Research shows significant user interest in models within the 10-30B parameter range for production use, as well as 3B models for mobile applications. To address these needs, the team has open-sourced Qwen2.5-3B, Qwen2.5-14B, and Qwen2.5-32B. Additionally, models like Qwen-Plus and Qwen-Turbo are available through API services on Alibaba Cloud Model Studio.

Qwen2.5 vs Qwen2

The Qwen2.5 series brings several key upgrades compared to the Qwen2 series:

  • Full-scale Open-source: In response to strong user demand for models in the 10-30B range for production and 3B models for mobile use, Qwen2.5 expands its open-source offerings beyond the four Qwen2 models (0.5B, 1.5B, 7B, and 72B). It adds two cost-effective, mid-sized models—Qwen2.5-14B and Qwen2.5-32B—as well as the mobile-optimized Qwen2.5-3B. These models are highly competitive, with Qwen2.5-32B outperforming Qwen2-72B and Qwen2.5-14B surpassing Qwen2-57B-A14B in comprehensive evaluations.
  • Larger and Higher Quality Pre-training Dataset: The pre-training dataset has expanded significantly, growing from 7 trillion tokens to 18 trillion tokens, enhancing the model’s training depth.
  • Knowledge Enhancement: Qwen2.5 demonstrates greater knowledge across various benchmarks. For instance, MMLU scores for Qwen2.5-7B and Qwen2.5-72B have risen to 74.2 and 86.1, compared to 70.3 and 84.2 for Qwen2 models, with substantial gains also observed in benchmarks like GPQA, MMLU-Pro, MMLU-redux, and ARC-c.
  • Coding Enhancement: Qwen2.5’s coding capabilities have improved significantly due to advancements in the Qwen2.5-Coder. Qwen2.5-72B-Instruct outperforms its predecessor on LiveCodeBench, MultiPL-E, and MBPP, with scores of 55.5, 75.1, and 88.2, respectively, versus 32.2, 69.2, and 80.2 for Qwen2-72B-Instruct.
  • Math Enhancement: Qwen2.5’s mathematical ability has also seen rapid improvement, with scores on the MATH benchmark increasing from 52.9 and 69.0 for Qwen2-7B/72B-Instruct to 75.5 and 83.1 for Qwen2.5 models.
  • Better Human Preference Alignment: Qwen2.5 generates responses more in line with human preferences. For example, the Arena-Hard score for Qwen2.5-72B-Instruct jumped from 48.1 to 81.2, and its MT-Bench score improved from 9.12 to 9.35, compared to Qwen2-72B-Instruct.
  • Other Core Capability Enhancements: Qwen2.5 models excel in instruction-following, long-text generation (expanding from 1K to over 8K tokens), understanding structured data, and producing structured outputs like JSON. They also demonstrate improved resilience to varied system prompts, enhancing role-playing and condition-setting for chatbots.

Qwen2.5 Model Card

The Qwen2.5 LLM models are released with seven open-sourced versions, ranging from 0.5B to 72B parameters. These models feature a context length of up to 128K tokens (131,072) and can generate text outputs of up to 8K tokens, making them highly effective for producing extensive and detailed content. Most of the models are available under the Apache 2.0 license, offering broad usage rights. However, the Qwen2.5-3B and Qwen2.5-72B models are governed by more specific licenses: the Qwen Research License and the Qwen License, respectively.

ModelsParamsNon-Emb ParamsLayersHeads (KV)Tie EmbeddingContext LengthGeneration LengthLicense
Qwen2.5-0.5B0.49B0.36B2414 / 2Yes32K8KApache 2.0
Qwen2.5-1.5B1.54B1.31B2812 / 2Yes32K8KApache 2.0
Qwen2.5-3B3.09B2.77B3616 / 2Yes32K8KQwen Research
Qwen2.5-7B7.61B6.53B2828 / 4No128K8KApache 2.0
Qwen2.5-14B14.7B13.1B4840 / 8No128K8KApache 2.0
Qwen2.5-32B32.5B31.0B6440 / 8No128K8KApache 2.0
Qwen2.5-72B72.7B70.0B8064 / 8No128K8KQwen
Qwen2.5 Model Family

Performance Overview

This section outlines the performance metrics of the Qwen2.5 models, focusing on both base and instruction-tuned models. These models are assessed across a wide range of domains, demonstrating their capabilities in natural language understanding, coding, mathematics, multilingual tasks, and more.

Qwen2.5 Base Language Model Evaluation

The evaluation of the Qwen2.5 base models primarily highlights their proficiency in natural language understanding, question answering, coding, mathematics, scientific knowledge, reasoning, and multilingual capabilities. Key evaluation datasets used include:

  • General Tasks: MMLU (5-shot), MMLU-Pro (5-shot), MMLU-redux (5-shot), BBH (3-shot), ARC-C (25-shot), TruthfulQA (0-shot), Winogrande (5-shot), HellaSwag (10-shot)
  • Math & Science Tasks: GPQA (5-shot), Theorem QA (5-shot), GSM8K (4-shot), MATH (4-shot)
  • Coding Tasks: HumanEval (0-shot), HumanEval+ (0-shot), MBPP (0-shot), MBPP+ (0-shot), MultiPL-E (0-shot) for Python, C++, Java, PHP, TypeScript, C#, Bash, JavaScript
  • Multilingual Tasks: Multi-Exam (M3Exam 5-shot, IndoMMLU 3-shot, ruMMLU 5-shot, mMMLU 5-shot); Multi-Understanding (BELEBELE 5-shot, XCOPA 5-shot, XWinograd 5-shot, XStoryCloze 0-shot, PAWS-X 5-shot); Multi-Mathematics (MGSM 8-shot); Multi-Translation (Flores-101 5-shot)

Qwen2.5-72B Performance

DatasetsLlama-3-70BMixtral-8x22BLlama-3-405BQwen2-72BQwen2.5-72B
General Tasks
MMLU79.577.885.284.286.1
MMLU-Pro52.851.661.655.758.1
MMLU-redux75.072.980.583.9
BBH81.078.985.982.486.3
ARC-C68.870.768.972.4
TruthfulQA45.651.054.860.4
WindoGrande85.385.086.785.183.9
HellaSwag88.088.787.387.6
Mathematics & Science Tasks
GPQA36.334.337.445.9
Theoremqa32.335.942.842.4
MATH42.541.753.850.962.1
MMLU-stem73.771.779.682.7
GSM8K77.683.789.089.091.5
Coding Tasks
HumanEval48.246.361.064.659.1
HumanEval+42.140.256.151.2
MBPP70.471.773.076.984.7
MBPP+58.458.163.969.2
MultiPL-E46.346.759.660.5
Multilingual Tasks
Multi-Exam70.063.576.678.7
Multi-Understanding79.977.780.789.6
Multi-Mathematics67.162.976.076.7
Multi-Translation38.023.337.839.0
Qwen2.5-72B Performance

The Qwen2.5-72B base model stands out by significantly outperforming other models in its category across a broad spectrum of tasks. It delivers results on par with Llama-3-405B, despite utilizing only one-fifth of the parameters. Additionally, when compared to its predecessor, Qwen2-72B, the Qwen2.5-72B demonstrates notable improvements in nearly all benchmark evaluations, with exceptional performance in general tasks, mathematics, and coding challenges.

Qwen2.5-14B/32B Performance

DatasetsQwen1.5-32BGemma2-27BYi-1.5-34BQwen2-57B-A14BQwen2.5-14BQwen2.5-32B
General Tasks
MMLU74.375.277.276.579.783.3
MMLU-pro44.149.148.343.051.255.1
MMLU-redux69.074.172.476.682.0
BBH66.874.976.467.078.284.5
ARC-C63.671.465.664.167.370.4
Truthfulqa57.440.153.957.758.457.8
Winogrande81.559.784.979.582.0
Hellaswag85.086.485.985.285.2
Mathematics & Science Tasks
GPQA30.834.937.434.332.848.0
Theoremqa28.835.840.033.543.044.1
MATH36.142.741.743.055.657.7
MMLU-stem66.571.072.669.876.480.9
GSM8K78.581.181.780.790.292.9
Coding Tasks
HumanEval43.354.946.353.056.758.5
HumanEval+40.246.340.246.351.252.4
MBPP64.275.765.571.976.784.5
MBPP+53.960.255.457.463.267.2
MultiPL-E38.548.039.549.853.559.4
Multilingual Tasks
Multi-Exam61.665.858.365.570.675.4
Multi-Understanding76.582.273.977.085.988.4
Multi-Mathematics56.161.649.362.368.573.7
Multi-Translation33.538.730.034.536.237.3
Qwen2.5-14B/32B Performance

The Qwen2.5-14B model delivers robust performance across a wide range of tasks, excelling in general tasks such as MMLU and BBH, with scores of 79.7 and 78.2, respectively, outperforming larger competitors. Similarly, Qwen2.5-32B demonstrates exceptional capabilities, frequently surpassing larger models of comparable sizes. It significantly outperforms its predecessor, Qwen1.5-32B, particularly in demanding areas like mathematics and coding, achieving impressive scores of 57.7 in MATH and 84.5 in MBPP.

Qwen2.5-7B Performance

DatasetsMistral-7BLlama3-8BGemma2-9BQwen2-7BQwen2.5-7B
#Non-emb Params7.0B7.0B8.2B6.5B6.5B
General Tasks
MMLU64.266.671.370.374.2
MMLU-pro30.935.444.740.145.0
MMLU-redux58.161.667.968.171.1
BBH56.157.768.262.370.4
ARC-C60.059.368.260.663.7
Trurhfulqa42.244.045.354.256.4
Winogrande78.477.479.577.075.9
Hellaswag83.382.181.980.780.2
Mathematics & Science Tasks
GPQA24.725.832.830.836.4
Theoremqa19.222.128.929.636.0
MATH10.220.537.743.549.8
MMLU-stem50.155.365.164.272.3
GSM8K36.255.370.780.285.4
Coding Tasks
HumanEval29.333.537.851.257.9
HumanEval+24.429.330.543.350.6
MBPP51.153.962.264.274.9
MBPP+40.944.450.651.962.9
MultiPL-E29.422.634.941.050.3
Multilingual Tasks
Multi-Exam47.152.361.259.259.4
Multi-Understanding63.368.678.372.079.3
Multi-Mathematics26.336.353.057.557.8
Multi-Translation23.331.936.531.532.4
Qwen2.5-7B Performance

The Qwen2.5-7B model outperforms both its predecessors and competitors across a range of benchmarks, despite utilizing fewer non-embedding parameters. It exhibits notable improvements in various tasks, scoring 74.2 on general benchmarks like MMLU, 49.8 on math-focused evaluations such as MATH, and 57.9 on coding tasks like HumanEval.

Qwen2.5-0.5B/1.5B/3B Performance

DatasetsQwen2-0.5BQwen2.5-0.5BQwen2-1.5BQwen2.5-1.5BGemma2-2.6BQwen2.5-3B
General Tasks
MMLU44.347.555.960.952.265.6
MMLU-pro14.715.721.628.523.034.6
MMLU-redux40.745.151.858.550.963.7
BBH18.220.336.545.141.956.3
ARC-C31.035.643.754.755.756.5
Trurhfulqa39.740.245.946.636.248.9
Winogrande56.956.365.065.071.571.1
Hellaswag49.152.167.067.974.674.6
Mathematics & Science Tasks
GPQA29.824.820.724.225.326.3
Theoremqa9.616.014.822.115.927.4
MATH11.219.521.635.018.342.6
MMLU-stem27.539.842.754.845.862.5
GSM8K36.441.646.968.530.379.1
Coding Tasks
HumanEval22.630.534.837.219.542.1
HumanEval+18.926.829.932.915.936.0
MBPP33.139.346.960.242.157.1
MBPP+27.633.837.649.633.649.4
MultiPL-E16.318.927.933.117.641.2
Multilingual Tasks
Multi-Exam29.430.843.147.938.154.6
Multi-Understanding40.441.050.765.146.876.6
Multi-Mathematics7.813.521.337.518.248.9
Multi-Translation14.115.323.825.026.929.3
Qwen2.5-0.5B/1.5B/3B Performance

For edge-side models, Qwen2.5-0.5B, 1.5B, and 3B demonstrate robust performance across nearly all benchmarks. Particularly, the Qwen2.5-0.5B model excels in math and coding tasks, surpassing the Gemma2-2.6B in these areas.

Instruction-tuned Model Evaluation

The evaluation of instruction-tuned models mainly focuses on the model performance of natural language understanding, general question answering, reasoning, coding, mathematics, instruction following, human alignment, etc.

The datasets for evaluation include:

General Tasks: MMLU-Pro, MMLU-redux

Math & Science Tasks: GPQA, GSM8K, MATH

Coding Tasks: HumanEval, MBPP, MultiPL-E, LiveCodeBench 2305-2409, LiveBench 0831

Instruction & Alignment Tasks: IFeval strict-prompt, Arena-Hard, AlignBench v1.1, MTbench

Qwen2.5-72B-Instruct Performance

DatasetsMistral-Large2 InstructLlama-3.1-70B-InstructLlama-3.1-405B-InstructQwen2-72B-InstructQwen2.5-72B-Instruct
MMLU-Pro69.466.473.364.471.1
MMLU-redux83.083.086.281.686.8
GPQA52.046.751.142.449.0
MATH69.968.073.869.083.1
GSM8K92.795.196.893.295.8
HumanEval92.180.589.086.086.6
MBPP80.084.284.580.288.2
MultiPL-E76.968.273.569.275.1
LiveCodeBench 2305-240942.232.141.632.255.5
LiveBench 083148.546.653.241.552.3
IFeval strict-prompt64.183.686.077.684.1
Arena-Hard73.155.769.348.181.2
AlignBench v1.17.695.945.958.158.16
MTbench8.618.799.089.129.35
Qwen2.5-72B-Instruct Performance

The Qwen2.5-72B-Instruct model delivers exceptional performance, even surpassing the larger Llama-3.1-405B in several critical tasks. Qwen2.5-72B-Instruct excels in mathematics (MATH: 83.1), coding (LiveCodeBench: 55.5), and chatting (Arena-Hard: 81.2). Compared to its base model Qwen2.5-72B and its predecessor Qwen2-72B-Instruct, the Qwen2.5-72B-Instruct showcases comprehensive improvements across all tasks.

Qwen-Turbo & Qwen2.5-14B-Instruct & Qwen2.5-32B-Instruct Performance

DatasetsQwen2-57B-A14B-InstructGemma2-27B-ITGPT4o-miniQwen-TurboQwen2.5-14B-InstructQwen2.5-32B-Instruct
MMLU-Pro52.855.563.164.863.769.0
MMLU-redux72.675.781.580.480.083.9
GPQA34.338.440.244.445.549.5
MATH49.154.470.281.080.083.1
GSM8K85.390.493.293.694.895.9
HumanEval79.978.788.486.683.588.4
MBPP70.981.085.780.282.084.0
MultiPL-E66.467.475.073.072.875.4
LiveCodeBench 2305-240922.540.743.142.651.2
LiveBench 083131.139.643.341.644.450.7
IFeval strict-prompt59.977.180.474.981.079.5
Arena-Hard17.857.574.968.468.374.5
AlignBench v1.17.027.227.817.997.947.93
MTbench8.559.108.868.889.20
Qwen-Turbo & Qwen2.5-14B-Instruct & Qwen2.5-32B-Instruct Performance

The Qwen2.5-32B-Instruct model demonstrates superior performance across most tasks when compared to other models of similar size. In comparison to GPT-4o-mini, our open-source model, Qwen2.5-14B-Instruct, along with our API model, Qwen-Turbo, also deliver competitive results across all benchmarks.

Qwen2.5-7B-Instruct Performance

DatasetsGemma2-9b-ITLlama3.1-8B-InstructQwen2-7B-InstructQwen2.5-7B-Instruct
MMLU-Pro52.148.344.156.3
MMLU-redux72.867.267.375.4
GPQA32.832.834.336.4
MATH44.351.952.975.5
GSM8K76.784.585.791.6
HumanEval68.972.679.984.8
MBPP74.969.667.279.2
MultiPL-E53.450.759.170.4
LiveCodeBench 2305-240918.98.323.928.7
LiveBench 083130.626.729.235.9
IFeval strict-prompt70.175.954.771.2
Arena-Hard41.627.825.052.0
AlignBench v1.17.054.757.137.33
MTbench8.498.238.268.75
Qwen2.5-7B-Instruct Performance

The Qwen2.5-7B-Instruct model significantly outperforms its competitors, Gemma2-9b-IT and Llama3.1-8B-Instruct, across all tasks except IFeval. Notably, Qwen2.5-7B-Instruct demonstrates clear advantages in mathematics (MATH: 75.5) and coding (HumanEval: 84.8).

Qwen2.5-3B-Instruct Performance

DatasetsGemma2-2B-ITPhi3.5-mini-InstructMiniCPM3-4BQwen2.5-3B-Instruct
Non-Emb Params2.0B3.6B4.0B2.8B
MMLU-Pro26.747.543.043.7
MMLU-redux51.967.759.964.4
GPQA29.327.231.330.3
MATH26.648.546.665.9
GSM8K63.286.281.186.7
HumanEval68.972.674.474.4
MBPP74.963.272.572.7
MultiPL-E30.547.249.160.2
LiveCodeBench 2305-24095.815.823.819.9
LiveBench 083120.127.427.626.8
IFeval strict-prompt51.052.168.458.2
Qwen2.5-3B-Instruct Performance

As for the edge-side instruction model, the Qwen2.5-3B-Instruct model has fewer parameters than both the Phi3.5-mini-Instruct and MiniCPM3-4B models. Despite this, it outperforms them in mathematics and coding tasks while delivering competitive results in language understanding.

Qwen2.5-0.5B/1.5B-Instruct Performance

DatasetsQwen2-0.5B-InstructQwen2.5-0.5B-InstructQwen2-1.5B-InstructQwen2.5-1.5B-Instruct
MMLU-Pro14.415.022.932.4
MMLU-redux12.924.141.250.7
GPQA23.729.821.229.8
MATH13.934.425.355.2
GSM8K40.149.661.673.2
HumanEval31.135.442.161.6
MBPP39.749.644.263.2
MultiPL-E20.828.538.550.4
LiveCodeBench 2305-24091.65.14.514.8
LiveBench 08317.412.612.418.8
IFeval strict-prompt14.627.929.042.5
Qwen2.5-0.5B/1.5B-Instruct Performance

Qwen2.5-1.5B-Instruct and Qwen2.5-0.5B-Instruct have seen large performance improvements over their previous versions, making them well-suited for edge-side applications in highly resource-constrained environments.

Performances on Multilingualism

To evaluate the multilingual performance of instruction-tuned models, we collect and extend benchmarks as follows:

  • IFEval (multilingual): We translate the examples from IFEval (English version) to construct multilingual IFEval examples after removing examples with language-specific contents (e.g., “start with letter A”). We collect 100 examples for each language among Arabic (ar), Spanish (es), French (fr), Indonesian (in), Japanese (ja), Korean (ko), Portuguese (pt), and Vietnamese (vi) languages. All examples are checked and post-edited (if neccessary) by paid volunteers.
  • Knowledge: We use 5 MMLU-like benchmarks (multi-choice) to testify the knowledge utilization ability of Qwen2.5 series models on multilingualism, including AMMLU (Arabic), JMMLU (Japanese), KMMLU (Korean), IndoMMLU (Indonesian), and TurkishMMLU (Turkish). Also, we present the performances on translated MMLU (i.e., okapi_MMLU, from English to multiple languages).
  • MGSM8K (extended): Aside from the examples in the original MGSM8K benchmark, we extend the language support with Arabic (ar), Korean (ko), Portuguese (pt), and Vietnamese (vi). We translate 250 examples (same as the other languages engaged in MGSM8K) into those 4 languages. All examples are also checked and post-edited (if necessary) by paid volunteers.
  • Cultural Nuances: We also use BLEnD, a benchmark aiming at testifying cultural nuances of LLMs, to testify LLMs from the Qwen2.5 series.
DatasetsQwen2-72B-InstructLlama3.1-70B-InstructQwen2.5-32B-InstructMistral-Large-Instruct-2407 (123B)GPT4o-miniQwen2.5-72B-Instruct
Instruction Following
IFEval (multilingual)79.6980.4782.6882.6985.0386.98
Knowledge
AMMLU (Arabic)68.8570.0870.4469.2469.7372.44
JMMLU (Japanese)77.3773.8976.5575.7773.7480.56
KMMLU (Korean)57.0453.2360.7556.4256.7761.96
IndoMMLU (Indonesian)66.3167.5066.4263.2167.7569.25
TurkishMMLU (Turkish)69.2266.8972.4164.7871.1976.12
okapi MMLU (translated)77.8476.4977.1678.3773.4479.97
Math Reasoning
MGSM8K (extended)82.7273.3187.1589.0187.3688.16
Cultural Nuances
BLEnD25.9030.4927.8833.4735.9132.48
Performances on Multilingualism
DatasetsQwen2-7B-InstructLlama3.1-8B-InstructQwen2.5-7B-InstructGemma-2-9B-InstructMistral-Nemo-Instruct-2407 (12B)Qwen2.5-14B-Instruct
Instruction Following
IFEval (multilingual)51.4360.6874.8777.4764.5977.08
Knowledge
AMMLU (Arabic)54.8754.2859.7860.2653.9266.81
JMMLU (Japanese)57.7153.2661.8864.5955.1772.78
KMMLU (Korean)43.9642.2846.5946.2442.2259.71
IndoMMLU (Indonesian)54.0553.9256.4261.7350.7665.09
TurkishMMLU (Turkish)49.2745.6154.2855.4434.4466.85
okapi MMLU (translated)60.4755.1866.9846.7259.6572.12
Math Reasoning
MGSM8K (extended)56.1366.0566.1178.3754.7582.27
Cultural Nuances
BLEnD22.4919.4723.6628.3126.6126.99
Performances on Multilingualism

Demo

Generating JSON Output

Structured Data Understanding

Long Text Generation

Links