How are Indian firms training LLMs? | Explained | Current Affairs | Vision IAS

Upgrade to Premium Today

Start Now
MENU
Home
Quick Links

High-quality MCQs and Mains Answer Writing to sharpen skills and reinforce learning every day.

Watch explainer and thematic concept-building videos under initiatives like Deep Dive, Master Classes, etc., on important UPSC topics.

A short, intensive, and exam-focused programme, insights from the Economic Survey, Union Budget, and UPSC current affairs.

ESC

Daily News Summary

Get concise and efficient summaries of key articles from prominent newspapers. Our daily news digest ensures quick reading and easy understanding, helping you stay informed about important events and developments without spending hours going through full articles. Perfect for focused and timely updates.

News Summary

Sun Mon Tue Wed Thu Fri Sat

How are Indian firms training LLMs? | Explained

26 Feb 2026
2 min

AI Impact Summit and Sarvam AI's Contribution

At the AI Impact Summit held in New Delhi, Sarvam AI, a Bengaluru-based startup, introduced two Large Language Models (LLMs) trained on 35 billion and 105 billion parameters. These models are designed to be less power- and compute-intensive while improving performance in Indian languages compared to other models.

Challenges for Indian Language LLMs

  • Data Availability:
    • Indian languages are underrepresented in internet data, creating challenges in training LLMs effectively.
  • Capital and Resource Scarcity:
    • Training LLMs requires significant financial and computational resources, which are limited in India.

Government Support and Initiatives

  • The IndiaAI Mission has subsidized LLM training, commissioning over 36,000 GPUs to support domestic AI development.
  • The government provided Sarvam with 4,096 GPUs, with an estimated subsidy of nearly ₹100 crore.
  • The Ministry of Electronics and Information Technology encourages domestic LLMs to enhance the Indian AI ecosystem.

Breakthroughs and Innovations

  • Mixture of Experts (MoE) Architecture:
    • This architecture allows the activation of only a fraction of parameters, enhancing efficiency and reducing computational needs.

Future Directions

Sarvam aims to enhance its model's depth and performance with future investments. The focus remains on accuracy and efficiency within the Indian context before expanding to larger models.

Other Indian LLM Developments

  • BharatGen, incubated by IIT Bombay, developed a multilingual 17 billion parameter model for applications in education and healthcare.

Explore Related Content

Discover more articles, videos, and terms related to this topic

RELATED TERMS

3

BharatGen

India's first multi-modal Large Language Model (LLM), a significant development in the country's AI journey. Understanding its capabilities and implications for India's technological sovereignty is important for UPSC.

GPUs (Graphics Processing Units)

Specialized electronic circuits designed to rapidly manipulate and alter memory to accelerate the creation of images in a frame buffer intended for output to a display device. In the context of AI, GPUs are crucial for training and running complex AI models due to their parallel processing capabilities.

IndiaAI Mission

A government mission focused on developing and deploying Artificial Intelligence in India, aiming to foster innovation, create a skilled workforce, and leverage AI for societal benefit. YUVA AI for ALL is an initiative under this mission.

Title is required. Maximum 500 characters.

Search Notes

Filter Notes

Loading your notes...
Searching your notes...
Loading more notes...
You've reached the end of your notes

No notes yet

Create your first note to get started.

No notes found

Try adjusting your search criteria or clear the search.

Saving...
Saved

Please select a subject.

Referenced Articles

linked

No references added yet