Indian AI startup Sarvam AI has launched Sarvam-1, the first LLM optimised specifically for Indian languages.
Developed with 2 billion parameters, Sarvam-1 supports 10 major Indian languages—Bengali, Gujarati, Hindi, Kannada, Malayalam, Marathi, Oriya, Punjabi, Tamil, and Telugu—alongside English.
Despite its relatively smaller size, Sarvam-1 shows strong performance in Indic language tasks, outperforming larger models like Gemma-2 and Llama-3 on benchmarks such as MMLU, ARC-Challenge, and IndicGenBench. It also offers faster inference speeds—4 to 6 times faster—making it suitable for deployment on edge devices.
For instance, on the TriviaQA benchmark, Sarvam-1 achieved an accuracy of 86.11 across Indic languages, significantly surpassing Llama-3.1 8B’s score of 61.47.
Sarvam-1’s performance on the IndicGenBench, which tests cross-lingual tasks such as summarization, translation, and question answering, also stood out. It achieved an average chrF++ score of 46.81 on Flores, a dataset for English-to-Indic translation, surpassing the larger Llama-3.1 8B model.
The model bridges the gap for Indian language speakers by offering advanced …