What Is AI Regulation? A Global Overview for 2026
A clear, accessible overview of AI regulation around the world in 2026—the EU AI Act, US executive orders and state laws, China's approach, and what builders and business leaders need to know.
View all what is ai depths →Depth ladder for this topic:
Why AI Regulation Matters Now
AI systems are making decisions that affect people’s lives—who gets a loan, who gets hired, what medical treatment is recommended, what content people see. As these systems become more powerful and pervasive, governments worldwide have moved from debating whether to regulate AI to debating how.
If you build, deploy, or use AI systems, regulation affects you. Not in some abstract future—now. The EU AI Act is being enforced. China has active AI regulations. The US has a patchwork of executive orders and state laws. Understanding this landscape isn’t optional for anyone working with AI professionally.
The EU AI Act: The Most Comprehensive Framework
The EU AI Act, which entered into force in August 2024 with phased implementation through 2026, is the world’s first comprehensive AI law. It establishes a risk-based framework that classifies AI systems into four tiers.
Unacceptable Risk (Banned)
These AI applications are prohibited entirely in the EU:
- Social scoring by governments (rating citizens based on behavior)
- Real-time biometric identification in public spaces (with narrow law enforcement exceptions)
- Emotion recognition in workplaces and educational institutions
- Manipulation techniques that exploit vulnerabilities (targeting children, people with disabilities)
- Predictive policing based solely on profiling
High Risk (Heavily Regulated)
AI systems in these domains face strict requirements:
- Employment: AI for hiring, promotion, termination decisions
- Education: AI for grading, admissions, learning assessment
- Critical infrastructure: AI controlling energy, transport, water systems
- Law enforcement: AI used in criminal justice (excluding banned uses)
- Credit scoring and insurance: AI that affects access to financial services
- Immigration and border control: AI used in visa and asylum decisions
Requirements for high-risk systems:
- Risk management systems with ongoing monitoring
- Data governance and quality requirements
- Technical documentation and record-keeping
- Transparency: users must know they’re interacting with AI
- Human oversight mechanisms
- Accuracy, robustness, and cybersecurity standards
- Conformity assessments before deployment
Limited Risk (Transparency Requirements)
AI systems like chatbots and deepfake generators must disclose that content is AI-generated. Users must be informed when they’re interacting with AI rather than a human.
Minimal Risk (No Specific Requirements)
Most AI systems—spam filters, recommendation engines, search algorithms—fall here and face no additional requirements beyond existing law.
General-Purpose AI Models (GPAI)
A separate category for foundation models and LLMs:
- All GPAI: Must provide technical documentation, comply with copyright law, and publish training data summaries
- Systemic risk GPAI (models trained with >10²⁵ FLOPs): Additional requirements including adversarial testing, incident reporting, and cybersecurity measures
Penalties
Non-compliance can result in fines up to €35 million or 7% of global annual revenue—whichever is higher. For context, that’s more severe than GDPR fines.
Practical Impact
If you sell AI products or services to EU customers:
- Classify your AI systems by risk level
- Implement required safeguards for high-risk systems
- Ensure transparency for limited-risk systems
- Document your AI systems’ training data, architecture, and testing
- Establish a compliance process and designate responsible personnel
The United States: Patchwork Approach
The US has no comprehensive federal AI law. Instead, regulation comes from multiple sources.
Federal Level
Executive Orders: The Biden administration’s October 2023 Executive Order on AI Safety established requirements for:
- Safety testing and reporting for powerful AI models
- Standards for AI safety and security (via NIST)
- Privacy protections in AI systems
- Equity and civil rights in AI use
- Consumer protection from AI-enabled fraud
The practical enforcement has been through existing agencies:
- FTC: Enforcement against deceptive AI practices, algorithmic discrimination
- EEOC: Guidance on AI in employment decisions
- HHS/FDA: Oversight of AI in healthcare and medical devices
- SEC: Scrutiny of AI claims in financial services and marketing
State Laws
States have been more active:
- Colorado AI Act (effective 2026): Requires impact assessments for “high-risk” AI decisions in employment, finance, housing, insurance, and healthcare
- California (multiple bills): Disclosure requirements for AI-generated content, restrictions on AI in hiring, proposed safety requirements for large models
- Illinois: Biometric information privacy (BIPA) affecting facial recognition
- New York City: Local Law 144 requires bias audits for AI used in hiring
- Texas, Virginia, Connecticut: Various AI transparency and consumer protection laws
Sector-Specific Regulation
Some US industries have specific AI requirements:
- Healthcare: FDA regulates AI/ML-based medical devices (over 500 approved to date)
- Financial services: Fair lending laws apply to AI credit decisions; model risk management guidelines from OCC/Fed
- Autonomous vehicles: State-by-state regulations for self-driving cars
China: Comprehensive and Centralized
China has taken a proactive, sector-specific approach:
- Algorithmic Recommendation Regulation (2022): Requires transparency in recommendation algorithms, opt-out mechanisms, and protections against addiction
- Deep Synthesis Regulation (2023): Governs deepfakes and AI-generated content. Requires labeling and traceability.
- Generative AI Measures (2023): Requires registration of generative AI services, content moderation, training data compliance, and truthfulness requirements
- AI Safety Governance Framework (2024): Comprehensive risk classification similar to the EU approach but with Chinese characteristics—stronger content control requirements, social stability considerations
Key differences from Western approaches:
- Content requirements align with Chinese political values
- Stronger government access to AI system internals
- Faster regulatory iteration (regulations are updated frequently)
- Mandatory filing/registration for AI services serving the public
Other Notable Jurisdictions
United Kingdom
The UK has taken a “pro-innovation” approach:
- No comprehensive AI law (deliberately)
- Existing regulators (FCA, Ofcom, CMA, ICO) apply AI-specific guidance within their domains
- AI Safety Institute conducts evaluations of frontier models
- Focus on voluntary commitments and international standards
Canada
- Artificial Intelligence and Data Act (AIDA): Part of broader digital charter legislation. Establishes requirements for “high-impact” AI systems including impact assessments and transparency.
- Aligned with but distinct from the EU approach
Brazil
- AI Regulation Bill: Moving through congress, establishing risk-based framework similar to EU approach
- Strong focus on non-discrimination and algorithmic transparency
Japan
- Light-touch, guidelines-based approach
- Focus on AI governance through industry self-regulation
- Strong emphasis on AI safety research
India
- No comprehensive AI regulation yet
- Advisory framework for “responsible AI”
- Sector-specific rules emerging (financial services, healthcare)
What This Means for Builders
If You Sell Globally
You effectively need to comply with the strictest applicable regulation—which usually means the EU AI Act. Building to EU standards gives you a baseline that satisfies most other jurisdictions.
Practical Compliance Steps
-
Inventory your AI systems: What AI do you build or use? What decisions does it influence?
-
Classify by risk level: Using the EU framework as a baseline, determine which systems are high-risk.
-
Document everything: Training data sources, model architecture, testing results, known limitations. This is required by the EU AI Act and good practice everywhere.
-
Implement human oversight: For high-risk systems, ensure humans can understand, monitor, and override AI decisions.
-
Test for bias: Regularly evaluate your systems for discriminatory outcomes across protected characteristics.
-
Establish transparency: Tell users when they’re interacting with AI. Explain how AI decisions are made.
-
Plan for incidents: Have a process for detecting, reporting, and addressing AI failures or harms.
-
Stay current: AI regulation is evolving rapidly. Assign someone to track changes.
For Startups
Compliance can feel overwhelming for small teams. Focus on:
- Documentation from day one (much harder to retrofit)
- Bias testing as part of your development process
- Transparency as a default (not something to add later)
- Understanding which regulations apply to your specific product and market
For Enterprise
Larger organizations should:
- Establish an AI governance committee or function
- Create internal AI policies aligned with applicable regulations
- Implement AI risk management frameworks (NIST AI RMF is a good starting point)
- Train employees on responsible AI use
- Engage with regulatory bodies and industry groups
The Trajectory
AI regulation is converging on several common principles across jurisdictions:
- Risk-based: Higher-stakes applications face stricter requirements
- Transparency: Users should know when AI is involved in decisions affecting them
- Accountability: Someone must be responsible for AI system outcomes
- Non-discrimination: AI systems shouldn’t produce discriminatory outcomes
- Safety: AI systems should be tested and monitored for harmful behavior
The details vary—the EU is prescriptive, the US is patchwork, the UK is principles-based, China is centralized—but the direction is clear. Building AI responsibly isn’t just ethically right; it’s increasingly legally required.
The organizations that treat compliance as an opportunity to build better, more trustworthy AI—rather than a burden to minimize—will have a lasting competitive advantage.
Simplify
← Narrow AI vs. General AI: What's the Real Difference?
Go deeper
AI Safety: What It Is and Why It Matters →
Related reads
Stay ahead of the AI curve
Weekly insights on AI — explained at the level that's right for you. No hype, no jargon, just what matters.
No spam. Unsubscribe anytime. We respect your inbox.