Browse Models
The simplest way to self-host Qwen 2.5 Math 72B. Launch a dedicated cloud GPU server running Lab Station OS to download and serve the model using any compatible app or framework.
Download model weights for local inference. Must be used with a compatible app, notebook, or codebase. May run slowly, or not work at all, depending on your system resources, particularly GPU(s) and available VRAM.
Qwen 2.5 Math 72B is a bilingual mathematical language model trained on over 1 trillion tokens, supporting both Chain-of-Thought and Tool-Integrated Reasoning. Available in base (1.5B/7B/72B) and instruction-tuned variants, it achieves 87.8% accuracy on MATH benchmarks, excelling in both academic and competitive mathematics.
Qwen 2.5 Math 72B represents a significant advancement in mathematical language models, building upon the previous Qwen2-Math series. Released in September 2024, this specialized large language model (LLM) is designed to solve complex mathematical problems in both English and Chinese languages. The model utilizes the Qwen2ForCausalLM architecture and is stored using the Safetensors format.
The model family includes several variants:
A key architectural innovation is the integration of both Chain-of-Thought (CoT) and Tool-Integrated Reasoning (TIR) methods, expanding upon its predecessor which only supported CoT for English mathematics problems. This dual-reasoning approach enables more sophisticated problem-solving capabilities across both languages.
The model was trained on the expanded Qwen Math Corpus v2, which contains over 1 trillion tokens - a significant increase from the previous version's 700 billion tokens. The dataset incorporates:
Qwen 2.5 Math 72B demonstrates exceptional capabilities in:
The model family, particularly the 72B instruction-tuned variant, has achieved state-of-the-art results across multiple benchmarks. Notable achievements include:
The model excels in various other benchmarks:
The model requires the Transformers library (version 4.37.0 or later) for implementation. The base model (Qwen2.5-Math-72B) is optimized for completion and few-shot learning tasks, while the instruction-tuned variant (Qwen2.5-Math-72B-Instruct) is designed for conversational interactions.
For optimal performance, the following parameter values are recommended when using different sampling methods:
The model is specifically designed for mathematical problem-solving and is not recommended for general-purpose tasks.