Indian Flag
Government Of India
A-
A
A+

sarvam-30B

Sarvam-30B is an advanced Mixture-of-Experts (MoE) model with 2.4B non-embedding active parameters, designed primarily for practical deployment. It combines strong reasoning, reliable coding ability, and best-in-class conversational quality across Indian languages. Sarvam-30B is built to run reliably in resource-constrained environments and can handle multilingual voice calls while performing tool calls.

About Model

Sarvam-30B is an advanced Mixture-of-Experts (MoE) model with 2.4B non-embedding active parameters, designed primarily for practical deployment. It combines strong reasoning, reliable coding ability, and best-in-class conversational quality across Indian languages. Sarvam-30B is built to run reliably in resource-constrained environments and can handle multilingual voice calls while performing tool calls. A major focus during training was the Indian context and languages, resulting in state-of-the-art performance across 22 Indian languages for its model size. Sarvam-30B is open-sourced under the Apache License. For more details, see our blog. Architecture The 30B MoE model is designed for throughput and memory efficiency. It uses 19 layers, a dense FFN intermediate_size of 8192, moe_intermediate_size of 1024, top-6 routing, grouped KV heads (num_key_value_heads=4), and an extremely high rope_theta (8e6) for long-context stability without RoPE scaling. It has 128 experts with a shared expert, a routed scaling factor of 2.5, and auxiliary-loss-free router balancing. The 30B model focuses on throughput and memory efficiency through fewer layers, grouped KV attention, and smaller experts. Read more at out blog - https://www.sarvam.ai/blogs/sovereign-models

sarvam-30B

Metadata Metadata

Apache 2.0

sarvamai

Mixture of Experts (MoE) Language Model

Transformers

Open

Sarvam AI

Sector Agnostic

04/03/26 15:35:41

59.92 GB

Activity Overview Activity Overview

  • Downloads1
  • Downloads 5
  • Views 56
  • File Size 59.92 GB

Tags Tags

  • MoE Model

License Control License Control

Apache 2.0

Version Control Version Control

FolderVersion 1(59.92 GB)
  • admin·3 day(s) ago
    • undefined
      __init__.py
    • undefined
      chat_template.jinja
    • application/json
      config.json
    • undefined
      configuration_sarvam_moe.py
    • application/json
      generation_config.json
    • undefined
      model-00001-of-00013.safetensors
    • undefined
      model-00002-of-00013.safetensors
    • undefined
      model-00003-of-00013.safetensors
    • undefined
      model-00004-of-00013.safetensors
    • undefined
      model-00005-of-00013.safetensors
    • more_horiz 14 more

More Models from Sarvam AI More Models from Sarvam AI

sarvam-30B
Sarvam-30B is an advanced Mixture-of-Experts (MoE) model with 2.4B non-embedding active parameters, designed primarily for practical deployment. It combines strong reasoning, reliable coding ability, and best-in-class conversational quality across Indian languages. Sarvam-30B is built to run reliably in resource-constrained environments and can handle multilingual voice calls while performing tool calls.
MoE Model
  • See Upvoters1
  • Downloads5
  • File Size59.92 GB
  • Views56
Updated 1 day(s) ago

SARVAM AI

sarvam-105b
Sarvam-105B is an advanced Mixture-of-Experts (MoE) model with 10.3B active parameters, designed for superior performance across a wide range of complex tasks. It is highly optimized for complex reasoning, with particular strength in agentic tasks, mathematics, and coding.
region:us
license:apache-2.0
bo
ks
sat
mni
doi
mai
kok
sd
ne
sa
ur
as
or
pa
ml
kn
gu
mr
te
ta
bn
hi
en
custom_code
conversational
Text Generation
sarvam_mla
Transformers
safetensors
  • See Upvoters0
  • Downloads10
  • File Size0
  • Views52
Updated 1 day(s) ago

SARVAM AI

sarvamM
Multilingual, hybrid-reasoning, text-only language model built on Mistral-Small
pa
Transformers
safetensors
mistral
Text Generation
conversational
en
bn
hi
kn
gu
mr
ml
or
ta
te
base_model:mistralai/Mistral-Small-3.1-24B-Base-2503
base_model:finetune:mistralai/Mistral-Small-3.1-24B-Base-2503
license:apache-2.0
autotrain_compatible
text-generation-inference
endpoints_compatible
region:us
  • See Upvoters1
  • Downloads48
  • File Size0
  • Views749
Updated 7 month(s) ago

SARVAM AI

sarvamtranslate
Translation model for 22 Indian Languages
kn
en
gu
gom
doi
brx
bn
as
Translation
image-text-to-text
gemma3
ks
hi
Transformers
safetensors
region:us
endpoints_compatible
text-generation-inference
license:gpl-3.0
base_model:finetune:google/gemma-3-4b-it
base_model:google/gemma-3-4b-it
ur
te
ta
sd
sat
sa
pa
or
ne
mr
mni
ml
mai
  • See Upvoters1
  • Downloads49
  • File Size0
  • Views821
Updated 8 month(s) ago

SARVAM AI

sarvam-1
India's first indic model, pretrained on 4 trillion tokens
  • See Upvoters1
  • Downloads196
  • File Size0
  • Views3,770
Updated 1 year(s) ago

SARVAM AI

shuka-v1
Multilingual audio to text model
audio-llms
  • See Upvoters2
  • Downloads81
  • File Size0
  • Views1,572
Updated 1 year(s) ago

SARVAM AI