Hercule is a cross-lingual evaluation model from the CIA Suite, fine-tuned on the INTEL dataset to assess multilingual LLMs using English references
Hercule is an advanced cross-lingual evaluation model introduced as part of the CIA Suite to assess multilingual Large Language Models (LLMs). It is fine-tuned on the INTEL dataset and demonstrates strong alignment with human evaluations, outperforming zero-shot proprietary models like GPT-4 on the RECON test set. Designed for low-resource languages, Hercule enables reference-based evaluation by scoring multilingual outputs using English reference responses. It provides structured feedback with a 1-5 scale scoring rubric and highlights the effectiveness of lightweight fine-tuning techniques like LoRA.
MIT
Sumanth Doddapaneni and Mohammed Safi Ur Rahman Khan and Dilip Venkatesh and Raj Dabre and Anoop Kunchukuttan and Mitesh M. Khapra
Evaluator Language model
N.A.
Open
Sector Agnostic
21/02/25 13:21:52
0
MIT
© 2026 - Copyright AIKosh. All rights reserved. This portal is developed by National e-Governance Division for AIKosh mission.