A 7-billion-parameter language model fine-tuned from LLaMA-2, designed to enhance reasoning capabilities in tasks such as reading comprehension, math problem-solving, and text summarization.
Orca-2-7B is a research-focused language model developed by Microsoft, fine-tuned from Meta's LLaMA-2 7B base model. It aims to improve the reasoning abilities of smaller language models by utilizing a synthetic dataset specifically curated for this purpose. The model excels in single-turn tasks, including reasoning over user-provided data, reading comprehension, mathematical problem-solving, and text summarization. Notably, Orca-2-7B has demonstrated performance comparable to much larger models, achieving significant improvements in complex zero-shot reasoning benchmarks. However, it is not optimized for chat applications and has not undergone reinforcement learning from human feedback (RLHF) or Direct Preference Optimization (DPO). Researchers are encouraged to fine-tune the model further for specific tasks or conversational use cases. The model is publicly available under the Microsoft Research License to support ongoing research in developing and aligning small language models.
Microsoft-research-license
Arindam Mitra and Luciano Del Corro and Shweti Mahajan and Andres Codas and Clarisse Simoes and Sahaj Agrawal and Xuxi Chen and Anastasia Razdaibiedina and Erik Jones and Kriti Aggarwal and Hamid Palangi and Guoqing Zheng and Corby Rosset and Hamed Khanpour and Ahmed Awadallah
Fine-Tuned LLaMA-2 Model
N.A.
Open
Sector Agnostic
12/03/25 06:34:53
0
Microsoft-research-license
© 2026 - Copyright AIKosh. All rights reserved. This portal is developed by National e-Governance Division for AIKosh mission.