BharatGPT Unveils Hanooman, a New Suite of Indic Generative AI Models

  • 24 Feb 2024

Why is it in the News?

Recently, the BharatGPT group, led by IIT Bombay along with seven other elite Indian engineering institutes announced that it would launch its first ChatGPT-like service next month.

What is Hanooman?

  • Hanooman is a series of large language models (LLMs) that can respond in 11 Indian languages like Hindi, Tamil, and Marathi, with plans to expand to more than 20 languages.
  • It is unveiled by Seetha Mahalaxmi Healthcare (SML) in partnership with the IIT Bombay-led BharatGPT ecosystem.
    • The BharatGPT group, which is backed by Reliance Industries.
  • Hanooman has been designed to work in four fields, including health care, governance, financial services, and education.
  •  According to BharatGPT, the series isn’t just a chatbot but It is a multimodal AI tool, which can generate text, speech, videos and more in multiple Indian languages.
  • One of the first customised versions is VizzhyGPT, an AI model fine-tuned for healthcare using reams of medical data.
    • The size of these AI models ranges from 1.5 billion to a whopping 40 billion parameters.

Are There Any Other Indian Language Models?

  • Apart from BharatGPT, a host of different startups like Sarvam and Krutrim, backed by prominent VC investors such as Lightspeed Venture Partners and billionaire Vinod Khosla’s fund, are also building AI models customised for India

What are Large Language Models (LLMs)?

  • Large language models use deep learning techniques to process large amounts of text.
  • They work by processing vast amounts of text, understanding the structure and meaning, and learning from it.
  • LLMs are ‘trained’ to identify meanings and relationships between words.
    • The greater the amount of training data a model is fed, the smarter it gets at understanding and producing text.
  • The training data is usually large datasets, such as Wikipedia, OpenWebText, and the Common Crawl Corpus.
  • These contain large amounts of text data, which the models use to understand and generate natural language.