AI Deployment in India: Navigating New Government Approval Requirements

AI News

< 1 Min Read

In-Short

  • India mandates government approval for AI technologies before public release.
  • AI developers must disclose ​potential unreliability in their technologies.
  • Deepfakes to be labelled with‍ unique identifiers to prevent misuse.
  • AI models must avoid bias ‍and⁣ not compromise electoral integrity.

Summary of India’s New AI Advisory

The Indian Ministry ‌of ⁣Electronics and Information Technology (MeitY) ‌has issued a new advisory that ​requires all AI technologies under development ⁢to⁢ obtain government​ permission prior to ‍public deployment. This move aims to regulate the burgeoning‌ AI sector⁢ and ‍address potential risks associated with⁢ AI systems.

Under the new guidelines, developers are​ also obligated to label their AI outputs,⁤ indicating‌ any possible errors or fallibility. A “consent ‌popup” ⁢feature is to be⁣ implemented to alert users about these potential defects. Additionally, the advisory calls for the labelling ‍of deepfakes⁣ with unique metadata to prevent ⁣their ⁣misuse.

Platforms hosting⁤ AI models, including large​ language‍ models (LLM), are ⁤instructed to ensure‌ their products⁣ are‌ free‌ from bias and ‌do not threaten the‍ integrity of electoral processes. The advisory‌ has been met ​with criticism from some industry experts, who argue ⁢that it could hinder innovation and progress in the AI field.

Developers‌ are expected ​to⁣ comply with the advisory within 15 ‌days of its release. Although⁢ the advisory is not legally binding at the moment, it reflects the government’s intentions and may foreshadow future legislation.⁤ IT Minister ‍Rajeev Chandrasekhar emphasized the importance of AI platforms taking full ⁣responsibility⁤ for‍ their products, even‌ during the‌ testing phase.

Further Information

For more ⁣detailed insights on India’s new AI ​regulations and ‍their implications, please refer to the original article.


Leave a Comment