India's Unique Approach to AI Governance
On November 5, 2025, the Ministry of Electronics and Information Technology (MeitY) released the India AI Governance Guidelines, establishing a comprehensive framework for artificial intelligence development and deployment across the nation.
Unlike the regulatory approaches adopted by the European Union and other jurisdictions, India has chosen a "techno-legal" framework that prioritizes innovation over restraint while embedding ethical safeguards.
The Seven Sutras: Core Principles
The governance framework is built upon seven foundational principles, referred to as "sutras":
1. Trust as the Foundation
Trust is positioned as essential for AI innovation, adoption, and risk mitigation across the entire AI value chain. This principle emphasizes:
- Transparency in AI system operations
- Verifiable claims about AI capabilities
- Building public confidence in AI technologies
2. People First
A human-centric approach requiring:
- Human oversight and control mechanisms
- User-friendly interfaces and explanations
- Protection of human rights and dignity
3. Innovation over Restraint
This principle distinguishes India's approach:
- Responsible innovation for socio-economic development
- Competitive positioning in global AI markets
- Avoiding overly restrictive regulations
4. Fairness & Equity
Ensuring AI systems deliver unbiased outcomes:
- Prevention of discrimination
- Equitable access to AI benefits
- Representation in training data
5. Accountability
Clear assignment of responsibility based on:
- Functions performed by AI systems
- Risk of harm assessment
- Due diligence requirements
6. Understandable by Design
Requiring transparency through:
- Clear explanations of AI operations
- Disclosure of AI system capabilities and limitations
- Accessible documentation
7. Safety, Resilience & Sustainability
Incorporating safeguards for:
- Risk minimization
- Anomaly detection
- Environmental responsibility
Institutional Framework
The guidelines establish three key institutions:
| Institution | Role |
|---|---|
| AI Governance Group (AIGG) | Strategic oversight and policy coordination |
| AI Safety Institute (AISI) | Technical validation and safety assessment |
| Technology & Policy Expert Committee (TPEC) | Expert advisory on emerging issues |
Addressing Specific Threats
The framework directly tackles emerging AI risks:
Deepfakes and Misinformation
- Mandatory content authentication
- Digital watermarks for AI-generated content
- Unique identifiers for traceability
Algorithmic Bias
- Audit requirements for high-risk AI systems
- Bias testing protocols
- Remediation mechanisms
National Security
- Sovereign infrastructure requirements
- Cross-border data flow restrictions
- Critical sector protections
Legal Integration
Rather than creating new legislation, the framework leverages existing laws:
- Information Technology Act, 2000
- Digital Personal Data Protection Act, 2023 (DPDPA)
- Consumer Protection Act
- Competition Act
Industry Response
The guidelines have been broadly welcomed by industry:
"This balanced approach allows India to remain competitive while ensuring responsible AI development." — NASSCOM
"The seven sutras provide clear principles that businesses can operationalize." — CII AI Task Force
Looking Ahead
The framework includes an action plan with:
- Short-term (6 months): Institution establishment
- Medium-term (18 months): Sector-specific guidelines
- Long-term (36 months): International alignment
India's AI governance approach positions the nation as a model for emerging economies seeking to balance innovation with accountability.







