Reasoning Model that gets a perfect score on IITJEE-25 Math Qs, surpassing o3-mini and o4-mini-low in math reasoning with just 14B params and under 16k context.
Fathom-R1-14B is a 14-billion-parameter reasoning language model derived from Deepseek-R1-Distilled-Qwen-14B, post-trained at an affordable cost of only $499, and achieving SOTA mathematical reasoning performance within a 16K context window. On the latest olympiad level exams: AIME-25 and HMMT-25, our model not only surpasses o3-mini-low, o1-mini and LightR1-14B(16k) at pass@1 scores (averaged over 64 runs) but also delivers performance rivaling closed-source o4-mini (low) w.r.t cons@64 — all while staying within a 16K context window. It achieves 52.71% Pass@1 accuracy on AIME2025 and 35.26% Pass@1 accuracy on HMMT25 (+7.2% and +5.2% improvement over the base model respectively). When provided with additional test-time compute in the form of cons@64, it achieves an impressive 76.7% accuracy on AIME2025 and 56.7% accuracy on HMMT25 (+13.4% and +6.7% improvement over the base model respectively). We perform supervised fine-tuning (SFT) on carefully curated datasets using a specific training approach, followed by model merging, achieving this performance at a total cost of just $499!
Model weights, training datasets, and fell $499 post training recipe are fully open-sourced!
Read more here - https://huggingface.co/FractalAIResearch/Fathom-R1-14B
MIT
FractalAIResearch
Large Language Models
PyTorch
Open
Science, Technology and Research
09/06/25 09:02:58
0
MIT
© 2026 - Copyright AIKosh. All rights reserved. This portal is developed by National e-Governance Division for AIKosh mission.