New open-source math model Light-R1-32B surpasses equivalent DeepSeek performance with only $1000 in training costs


Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


A team of researchers has introduced Light-R1-32B, a new open-source AI model optimized for solving advanced math problems, making it available on Hugging Face under a permissive Apache 2.0 license — free for enterprises and researchers to take, deploy, fine-tune or modify as they wish, even for commercial purposes.

The 32-billion parameter (number of model settings) model surpasses the performance of similarly sized (and even larger) open source models such as DeepSeek-R1-Distill-Llama-70B and DeepSeek-R1-Distill-Qwen-32B on third-party benchmark the American Invitational Mathematics Examination (AIME), which contains 15 math problems designed for extremely advanced students and has an allotted time limit of 3 hours for human users.

Developed by Liang Wen, Fenrui Xiao, Xin He, Yunke Cai, Qi An, Zhenyu Duan, Yimin Du, Junchen Liu, Lifu Tang, Xiaowei Lv, Haosheng Zou, Yongchao Deng, Shousheng Jia, and Xiangzheng Zhang, the model surpasses previous open-source alternatives on competitive math benchmarks.

Incredibly, the researchers completed the model’s training in fewer than six hours on 12 Nvidia H800 GPUs at an estimated total cost of $1,000. This makes Light-R1-32B one of the most accessible and practical approaches for developing high-performing math-specialized AI models. However, it’s important to remember the model was trained on a variant of Alibaba’s open source Qwen 2.5-32B-Instruct, which itself is presumed to have had much higher upfront training costs.

Alongside the model, the team has released its training datasets, training scripts, and evaluation tools, providing a transparent and accessible framework for building math-focused AI models.

The arrival of Light-R1-32B follows other similar efforts from rivals such as Microsoft with its Orca-Math series.

A new math king emerges

Light-R1-32B is designed to tackle complex mathematical reasoning, particularly on the AIME (American Invitational Mathematics Examination) benchmarks.

It was trained from Qwen2.5-32B-Instruct, starting from a model without long-chain-of-thought (COT) reasoning. The team applied curriculum-based supervised fine-tuning (SFT) and Direct Preference Optimization (DPO) to refine its problem-solving capabilities.

When evaluated, Light-R1-32B achieved 76.6 on AIME24 and 64.6 on AIME25, surpassing DeepSeek-R1-Distill-Qwen-32B, which scored 72.6 and 54.9, respectively.

This improvement suggests that the curriculum-based training approach effectively enhances mathematical reasoning, even when training from models that initially lack long COT.

Fair benchmarking

To ensure fair benchmarking, the team decontaminated training data against common reasoning benchmarks, including AIME24/25, MATH-500, and GPQA Diamond, preventing data leakage.

They also implemented difficulty-based response filtering using DeepScaleR-1.5B-Preview, ultimately forming a 76,000-example dataset for the first stage of supervised fine-tuning. A second, more challenging dataset of 3,000 examples further improved performance.

After training, the team merged multiple trained versions of Light-R1-32B, leading to additional gains. Notably, the model maintains strong generalization abilities on scientific reasoning tasks (GPQA), despite being math-specialized.

How enterprises can benefit

Light-R1-32B is released under the Apache License 2.0, a permissive open-source license that allows free use, modification, and commercial deployment without requiring derivative works to be open-sourced. T

his makes it an attractive option for enterprises, AI developers, and software engineers looking to integrate or customize the model for proprietary applications.

The license also includes a royalty-free, worldwide patent grant, reducing legal risks for businesses while discouraging patent disputes. Companies can freely deploy Light-R1-32B in commercial products, maintaining full control over their innovations while benefiting from an open and transparent AI ecosystem.

For CEOs, CTOs, and IT leaders, Apache 2.0 ensures cost efficiency and vendor independence, eliminating licensing fees and restrictive dependencies on proprietary AI solutions. AI developers and engineers gain the flexibility to fine-tune, integrate, and extend the model without limitations, making it ideal for specialized math reasoning, research, and enterprise AI applications. However, as the license provides no warranty or liability coverage, organizations should conduct their own security, compliance, and performance assessments before deploying Light-R1-32B in critical environments.

Transparency in low-cost training and optimization for math problem solving

The researchers emphasize that Light-R1-32B provides a validated, cost-effective way to train strong long-chain-of-thought models in specialized domains.

By sharing their methodology, training data, and code, they aim to lower the cost barriers for high-performance AI development.

Future work includes exploring reinforcement learning (RL) to enhance the model’s reasoning capabilities further.



Source link

Share

Latest Updates

Frequently Asked Questions

Related Articles

Google rolls out Nano Banana to Search, NotebookLM

Google’s advanced digital tool Nano Banana has been rolled out to Google Search...

Bombay HC slams deepfake abuse of Suniel Shetty’s persona as ‘depraved misuse of technology’

In a strongly worded order that addresses the growing threat of artificial intelligence-driven...

MIT study reveals chilling cognitive cost of ChatGPT on children: What can parents do?

As artificial intelligence becomes part of daily life, a new concern is emerging...

This new AI technique creates ‘digital twin’ consumers, and it could kill the traditional survey industry

A new research paper quietly published last week outlines a breakthrough method that...
sabung ayam online sabung ayam online judi bola sabung ayam online judi bola Judi Bola Sabung Ayam Online Live Casino Online Sabung Ayam Online Sabung Ayam Online Sabung Ayam Online Sabung Ayam Online Sabung Ayam Online Sabung Ayam Online sabung ayam online judi bola mahjong ways sabung ayam online judi bola mahjong ways mahjong ways sabung ayam online sv388 Sv388 judi bola judi bola judi bola judi bola JUARA303 Mahjong ways Judi Bola Judi Bola Sabung Ayam Online Live casino mahjong ways 2 sabung ayam online sabung ayam online mahjong ways mahjong ways mahjong ways live casino online sabung ayam online judi bola SV388 SBOBET88 judi bola judi bola judi bola judi bola judi bola https://himakom.fisip.ulm.ac.id/ SABUNG AYAM ONLINE MIX PARLAY SLOT GACOR JUDI BOLA SV388 LIVE CASINO LIVE CASINO ONLINE Judi Bola Online SABUNG AYAM ONLINE JUDI BOLA ONLINE LIVE CASINO ONLINE JUDI BOLA ONLINE LIVE CASINO ONLINE LIVE CASINO ONLINE sabung ayam online Portal SV388 SBOBET88 SABUNG AYAM ONLINE JUDI BOLA ONLINE CASINO ONLINE MAHJONG WAYS 2 sabung ayam online judi bola SABUNG AYAM ONLINE JUDI BOLA ONLINE Sabung Ayam Online JUDI BOLA Sabung Ayam Online JUDI BOLA SV388, WS168 & GA28 SBOBET88 SV388, WS168 & GA28 SBOBET88 SBOBET88 CASINO ONLINE SLOT GACOR Sabung Ayam Online judi bola judi bola judi bola judi bola --indomax77 judi bola online --indomax77 mix parlay --indomax77 situs mix parlay --indomax77 situs parlay --indomax77 sbobet --indomax77 sbobet88 --indomax77 situs bola --indomax77 situs judi bola --indomax77 agen bola --indomax77 agen judi bola --indomax77 agen mix parlay --indomax77 agen parlay --indomax77