Training of Large Language Models | 26 Feb 2026

Source: TH

Why in News? 

At the India-AI Impact Summit 2026, Sarvam AI unveiled two indigenous Large Language Models (LLMs), marking a significant step in India’s AI ecosystem. 

  • The models are designed to be less power- and compute-intensive while improving performance in Indian languages, highlighting India’s growing capability in training cost-efficient, locally relevant LLMs.

How are Large Language Models (LLMs) Trained?

  • LLM: A Large Language Model (LLM) is an artificial intelligence system built using transformer-based neural networks that can understand, interpret, and generate human language. 
  • Trained on massive datasets, LLMs use deep learning to identify patterns in text and learn how words and sentences relate to each other. 
  • Their performance depends on data quality, and they are further improved through fine-tuning for specific tasks such as translation, summarization, and question answering.

Training

  • Data Collection & Pre-processing: LLM training begins with collecting a massive corpus of text from sources like the internet, books, Wikipedia, and code repositories. 
    • The text is converted into tokens (small units of words or subwords) so machines can process it. 
    • The data is then cleaned to remove spam, duplicates, bias, and harmful content, ensuring the model learns accurate and high-quality language patterns.
  • Pre-training (Self-Supervised Learning): In the pre-training stage, the model learns language patterns using next-token prediction, where it predicts the next word in a sentence. 
    • Most modern LLMs use the Transformer architecture with self-attention, enabling them to understand relationships between distant words. 
    • This stage produces a base model that understands grammar, facts, and reasoning but cannot yet follow instructions effectively.
  • Supervised Fine-Tuning (Instruction Tuning): During supervised fine-tuning, the model is trained on curated prompt–response pairs created by human experts. 
    • This teaches the model how to respond to instructions, answer questions, and perform tasks like summarization. 
    • As a result, the model learns the format of conversations and becomes more useful for real-world applications.
  • Alignment using RLHF:  In the alignment stage, developers use RLHF (Reinforcement Learning from Human Feedback) to ensure the model produces safe, unbiased, and helpful responses. 
    • The model generates multiple answers, which humans rank based on quality and safety. 
    • A reward model learns these preferences, and the LLM is optimized to produce responses aligned with human values and ethical standards.
  • Limitations of Training LLMs in India: Indian AI models face several bottlenecks, including the scarcity of high-quality datasets in Indian languages, which affects accuracy and inclusivity. 
    • Many models consume more tokens by translating inputs into English to improve performance, increasing computational costs. 
    • Additionally, training LLMs in India remains challenging due to limited capital availability and the lack of immediate commercial use cases, which constrains large-scale investment in domestic AI development.

    Note: Early Large Language Models (LLMs) with hundreds of billions or even trillions of parameters processed queries by activating all parameters during inference, making them computationally expensive and resource-intensive.

    • The new models use Mixture of Experts (MoE), an AI model architecture where only a small subset of parameters (“experts”) is activated for each query, instead of using the entire model. 
    • This makes LLMs faster, cheaper, and more compute-efficient, enabling cost-effective deployment in resource-constrained environments like India.
    • Indian models such as Sarvam’s 105B parameter LLM and BharatGen’s multilingual 17B model use efficiency-focused approaches to support local languages and sectors like education and healthcare.

    IndiaAI Mission

    • Launched in March 2024 with an outlay of Rs 10,372 crore, the IndiaAI Mission aims to build a comprehensive AI ecosystem in India. 
    • Under the mission, the government has commissioned over 36,000 Graphics Processing Units (GPUs)  in Indian data centres and is expanding capacity by adding 20,000 more, targeting over 100,000 GPUs by the end of 2026.
      • Startups such as Sarvam AI have received subsidised access to compute infrastructure for AI training and inference. Sarvam AI was granted access to 4,096 GPUs from the common compute cluster, with subsidies amounting to nearly Rs 100 crore.
    • The mission also supports talent development for over 13,500 students, is establishing India Data and AI Labs, and promotes sovereign foundational models trained on Indian datasets, with financial assistance covering compute and related costs to boost open-source innovation and startup growth.

    Frequently Asked Questions (FAQs)

    1. What is a Large Language Model (LLM)?
      An LLM is a transformer-based AI system trained on massive datasets to understand and generate human language using deep learning.

    2. What is Mixture of Experts (MoE) architecture in AI?
      MoE activates only a subset of model parameters per query, reducing compute costs and improving efficiency in LLM deployment.

    3. What is the objective of the IndiaAI Mission?
      It aims to build a sovereign AI ecosystem through compute infrastructure, indigenous models, datasets, and talent development.

    4. Why are indigenous LLMs important for India?
      They improve Indian language support, ensure data sovereignty, and reduce dependence on foreign AI platforms.

    5. How does RLHF improve AI models?
      Reinforcement Learning from Human Feedback aligns AI outputs with human values, ensuring safety, accuracy, and ethical responses.

    UPSC Civil Services Examination Previous Year Question (PYQ)

    Prelims

    Q. With the present state of development, Artificial Intelligence can effectively do which of the following? (2020)

    1. Bring down electricity consumption in industrial units
    2. Create meaningful short stories and songs
    3. Disease diagnosis
    4. Text-to-Speech Conversion
    5. Wireless transmission of electrical energy

    Select the correct answer using the code given below:

    (a) 1, 2, 3 and 5 only
    (b) 1, 3 and 4 only
    (c) 2, 4 and 5 only
    (d) 1, 2, 3, 4 and 5

    Ans: (b)