Updates
GS Paper IIIEnvironmental Ecology

Training of Large Language Models (LLMs) by Indian Firms

LearnPro Editorial
26 Feb 2026
Updated 3 Mar 2026
8 min read
Share

Sarvam AI's LLMs: A Watershed Moment or Another Nascent Step?

On February 25, 2026, Bengaluru-based Sarvam AI announced its entry into the evolving field of indigenous Large Language Models (LLMs). With models trained on 35 billion and 105 billion parameters and a strong emphasis on multilingual functionality, this development signals India’s growing ambition to assert itself in the global AI race. However, beneath the milestone lies a complicated reality: an ecosystem simultaneously promising and precarious.

Notably, these models promise a decisive pivot toward accessibility in Indian regional languages. Given that over 90% of Indians reportedly prefer content in their native tongue, such a capability could democratize access to AI-driven services. However, the critical question remains: Will these models escape the commercial and infrastructural challenges that have historically constrained India's tech aspirations?

Institutional Framework and IndiaAI's Role

At the center of India’s AI ambitions lies the IndiaAI Mission, the flagship policy initiative envisioned to establish AI sovereignty for the nation. With its twin focus on building computational infrastructure and fostering indigenous research and development, the Mission provided part of the ecosystem that birthed Sarvam AI’s models. Consider the following achievements:

  • India now boasts a network of 38,000 GPUs, an essential backbone for AI development in the country.
  • The Mission collaborates with key institutes like IIT Bombay, which incubated BharatGen, a 17-billion-parameter multilingual model intended for education and healthcare.
  • Support is extended to AI-focused startups such as Gnani.ai, which advanced compact speech and text-to-speech models.

While these efforts are notable on paper, their impact will hinge on execution. Sarvam AI’s models claim to be open-source, marking potential accessibility gains. Yet, without stringent peer review and cross-institutional collaboration, the broader AI research community might hesitate to engage fully. The Ministry of Electronics and IT (MeitY), tasked with overseeing these models under the IndiaAI framework, must ensure that rhetoric translates into meaningful adoption.

Beyond the Hype: Ground-Level Challenges

Training LLMs is a resource-intensive undertaking. Sarvam AI’s models, while impressive in size, expose the structural roadblocks facing AI ambitions in India. Consider the following:

First, data scarcity. While Sarvam AI has purportedly curated high-quality datasets in Indian languages, the larger ecosystem is plagued by a lack of structured, high-volume regional language corpora. This is particularly problematic when over half of users engaging with digital platforms in Indian languages report subpar personalization.

Second, the costs of access. Training models like Sarvam AI's requires prolonged engagement with GPU clusters – operations that run into months and demand enormous capital. For instance, OpenAI reportedly incurred upwards of $5 million in direct training costs for GPT-3. Indian startups, without assured commercial returns, face significant sustainability risks. This highlights the urgency of subsidies or shared infrastructure under the IndiaAI framework.

Third, there’s the question of performance. While Sarvam AI’s focus on Indian languages appears mission-aligned, early outputs reveal performance inconsistencies compared to English benchmarks. This mirrors lessons from BharatGen and other indigenous models, many of which have shown marked drops in accuracy during application-level testing for non-English users.

At a deeper level, the reliance on translation-centric workflows—where input sentences in regional languages are often translated to English before processing—adds latency and compromises interpretability. Worse, such approaches raise costs by increasing token requirements. The dream of seamless multilingual AI remains just that: a dream, unless structural fixes are executed at scale.

Structural Faultlines: Capital and Coordination

India’s AI revolution also risks being throttled not by ambition but by economics and federal design. A telling example can be found in Sarvam AI’s reliance on GPU-based training. While India’s 38,000 GPUs sound momentous, the reality is starkly different when one considers availability per startup and region. Moreover, such resources lack the scale of those housed by global hyperscalers like Microsoft or Google.

The funding gap compounds this further. India’s allocation to AI initiatives under multiple budgetary heads exceeds ₹4,500 crore collectively, but a significant portion remains tied to short-term project milestones rather than large-scale foundational investments. In comparison, China reportedly invested nearly $1.6 billion in just one of its regional AI hubs in Guangzhou—a scale India has not achieved yet.

Inter-ministerial coordination presents another challenge. While MeitY steers the IndiaAI Mission, the Ministry of Skills Development and Entrepreneurship (MSDE) lags behind in creating a trained workforce dedicated to AI. Absence of synergy between these institutions may create bottlenecks in expanding AI adoption at the industry level.

International Lens: Lessons (and Warnings) from China

China offers a contrasting model of execution. Through its state-backed research programs and private partnerships, China has trained LLMs like Wu Dao 2.0, boasting a staggering 1.75 trillion parameters. Critical to this success has been seamless data-sharing agreements, high-efficiency national supercomputing centres, and multi-billion-dollar investments in AI clusters.

While India’s commitment to sovereignty over compute and data has merit, its go-it-alone approach risks inefficiencies. Beijing’s high reliance on closed data ecosystems and single-party coordination may hold lessons for India: if we cannot match the resource intensity of China, can we at least systematize partnerships between institutions like Sarvam AI, IISc, and global tech players?

Steps Ahead: What Success Will Demand

The real success of Sarvam AI—or any indigenous LLM project—will depend on three critical areas. First, scaling access to regional language datasets. This demands nothing short of a curated public-private partnership, pooling resources and knowledge in corpus generation.

Second, IndiaAI must consider moving beyond GPU-scale debates toward creating affordable national AI infrastructure. Shared cloud platforms, akin to Singapore’s National AI Program, would mitigate resource bottlenecks and reduce operating costs for domestic startups.

Finally, metrics for success must be set and independently audited. Parameters like not just adoption rates but reductions in language processing times and domain adaptability should be monitored to ensure long-term benefits.

📝 Prelims Practice
  1. Which of the following methodologies is used to align LLMs with human intent?
    • a) Transfer Learning
    • b) Reinforcement Learning from Human Feedback
    • c) Gradient Descent Clustering
    • d) Token Augmentation
  2. What is the primary bottleneck when training LLMs such as Sarvam AI's models?
    • a) Lack of GPU-based training facilities
    • b) High computation costs
    • c) Scarcity of regional language corpus
    • d) Absence of open-source tools
✍ Mains Practice Question
Critically evaluate whether India’s sovereign AI ambitions, as seen in the development of indigenous LLMs, adequately address the structural limitations in data accessibility, compute infrastructure, and inter-institutional coordination.
250 Words15 Marks

Practice Questions for UPSC

Prelims Practice Questions

📝 Prelims Practice
Consider the following statements about barriers to building effective Indian-language LLMs:
  1. Scarcity of structured, high-volume regional-language corpora can weaken personalization and model usefulness for non-English users.
  2. Translation-centric pipelines can increase latency and token usage, potentially raising costs while complicating interpretability.
  3. A larger national GPU count automatically ensures adequate compute availability for every startup and region.

Which of the above statements is/are correct?

  • a1 and 2 only
  • b2 and 3 only
  • c1 and 3 only
  • d1, 2 and 3
Answer: (a)
📝 Prelims Practice
Consider the following statements about India’s institutional support for indigenous LLMs as described in the article:
  1. The IndiaAI Mission aims at AI sovereignty with emphasis on computational infrastructure and indigenous research and development.
  2. IIT Bombay incubated BharatGen, a multilingual model intended for sectors like education and healthcare.
  3. The article concludes that open-source claims alone are sufficient to guarantee broad research community participation.

Which of the above statements is/are correct?

  • a1 and 2 only
  • b2 and 3 only
  • c1 and 3 only
  • d1, 2 and 3
Answer: (a)
✍ Mains Practice Question
Critically examine India’s push for indigenous Large Language Models in light of the IndiaAI Mission. Discuss the structural constraints highlighted in the article—data scarcity, compute costs, performance gaps in Indian languages, and coordination/funding challenges—and suggest measures to ensure sustainable, scalable adoption. (250 words)
250 Words15 Marks

Frequently Asked Questions

Why is multilingual capability in indigenous LLMs considered important for India’s AI adoption?

The article links multilingual capability to accessibility because a large majority of Indians prefer content in their native language. If LLMs perform well across regional languages, they can widen access to AI services beyond English-first users and reduce exclusion in digital public services and markets.

What role does the IndiaAI Mission play in enabling indigenous LLM development, and what are its key focus areas?

The IndiaAI Mission is described as a flagship initiative aimed at building AI sovereignty through computational infrastructure and support for indigenous R&D. It has helped create enabling conditions such as a national GPU network and collaborations with institutions and startups, but outcomes depend heavily on execution and adoption.

What institutional and governance issues does the article highlight around open-source claims and adoption of LLMs?

Even if models are claimed to be open-source, the article notes that without stringent peer review and cross-institutional collaboration, researchers may hesitate to engage. It also places responsibility on MeitY, which oversees the models under the IndiaAI framework, to ensure that policy intent translates into meaningful adoption.

Why does the article consider data scarcity a structural bottleneck for Indian-language LLMs?

The article points to a broader lack of structured, high-volume corpora in regional languages, which limits training quality and downstream performance. It connects this to user experience by noting that many Indian-language users report subpar personalization, suggesting the dataset ecosystem is not yet mature.

How do translation-centric workflows affect multilingual LLM performance and costs according to the article?

The article argues that translating regional-language inputs into English before processing increases latency and reduces interpretability. It also raises costs by increasing token requirements, meaning the approach can be both slower and more expensive while still failing to deliver seamless multilingual performance.

Source: LearnPro Editorial | Environmental Ecology | Published: 26 February 2026 | Last updated: 3 March 2026

Share
About LearnPro Editorial Standards

LearnPro editorial content is researched and reviewed by subject matter experts with backgrounds in civil services preparation. Our articles draw from official government sources, NCERT textbooks, standard reference materials, and reputed publications including The Hindu, Indian Express, and PIB.

Content is regularly updated to reflect the latest syllabus changes, exam patterns, and current developments. For corrections or feedback, contact us at admin@learnpro.in.

This Topic Is Part Of

Related Posts

Science and Technology

Missile Defence Systems

Context The renewed hostilities between the United States-led coalition (including Israel and United Arab Emirates) and Iran have tested a newly integrated regional air and missile defence network in West Asia. What is a missile defence system? Missile defence refers to an integrated military system designed to detect, track, intercept, and destroy incoming missiles before they reach their intended targets, thereby protecting civilian populations, military installations, and critical infrastruct

2 Mar 2026Read More
International Relations

US-Israel-Iran War

Syllabus: GS2/International Relations Context More About the News Background of the Current Escalation Global Implications Impact on India Way Forward for India About West Asia & Its Significance To Global Politics Source: IE

2 Mar 2026Read More
Polity

Securities and Exchange Board of India (SEBI) on Market Manipulators

Context The Securities and Exchange Board of India (SEBI) will enhance surveillance and enforcement on market manipulators and cyber fraudsters through technology and use Artificial Intelligence (AI). Securities and Exchange Board of India (SEBI) It is the regulatory authority for the securities and capital markets in India. It was established in 1988 and given statutory powers through the SEBI Act of 1992.

2 Mar 2026Read More
Polity

18 February 2026 as a Current Affairs Prompt: How to Convert a Date into UPSC Prelims-Grade Facts (Acts, Rules, Notifications, Institutions)

A bare date like “18-February-2026” is not a defensible current-affairs topic unless it is anchored to a primary instrument such as a Gazette notification, regulator circular, court judgment, or a Bill/Act. The exam-relevant task is to convert the date into verifiable identifiers—issuing authority, legal basis (Act/Rules/Sections), instrument number, effective date, and thresholds—because UPSC frames MCQs around precisely these hard edges. The central thesis: the difference between narrative awareness and Prelims accuracy is source hierarchy discipline.

2 Mar 2026Read More

Enhance Your UPSC Preparation

Study tools, daily current affairs analysis, and personalized study plans for Civil Services aspirants.

Try LearnPro AI Free

Our Courses

72+ Batches

Our Courses
Contact Us