It is a multilingual, sequence-to-sequence pre-trained model fine-tuned on the IndicBARTSS checkpoint for headline generation and summarization tasks across 11 Indian languages.
It is a sequence-to-sequence pre-trained model focusing on 11 Indian languages: Hindi, Marathi, Punjabi, Tamil, Telugu, Bengali, Gujarati etc. It is fine-tuned on the IndicBARTSS checkpoint and can be used for tasks like summarization, headline generation, and other related applications in Indian languages. The model is smaller than mBART and mT5, making it computationally efficient for fine-tuning and decoding. Trained on a large corpus of 1.316 million paragraphs and 5.9 million unique tokens, it supports each language in its own script, eliminating the need for Devanagari script mapping.
MIT
Aman Kumar, Himani Shrotriya, Prachi Sahu, Raj Dabre, Ratish Puduppully, Anoop Kunchukuttan, Amogh Mishra, Mitesh M. Khapra, Pratyush Kumar
Text Summarization
N.A.
Open
Sector Agnostic
21/02/25 13:20:55
0
MIT
© 2026 - Copyright AIKosh. All rights reserved. This portal is developed by National e-Governance Division for AIKosh mission.