The most important developments in AI, explained simply. Updated regularly.
Last updated: 2026-04-03
Model Release
Google Releases Gemma 4 Open Models Under Apache 2.0
Google DeepMind releases Gemma 4, a family of four open models (E2B, E4B, 26B MoE, 31B Dense) built on Gemini 3 research. The 31B Dense ranks #3 among open models on the Arena AI leaderboard, scoring 89.2% on AIME 2026 and 80.0% on LiveCodeBench v6. First Gemma release under a fully OSI-approved Apache 2.0 license.
California Issues First-of-Its-Kind AI Executive Order
Governor Newsom signs an executive order setting new AI procurement standards for state contracts, directing the California Department of Technology to create first-in-the-nation recommendations for watermarking AI-generated content, and enabling the state to separate its AI procurement process from the federal government.
Google releases Gemini 3.1 Flash Live, a real-time multimodal voice model for low-latency conversations with audio, video, and tool use. Supports 90+ languages and rolls out across 200+ countries via Search Live and the Gemini API.
The European Parliament votes 569-45 to amend the AI Act, banning AI 'nudifier' systems that generate non-consensual intimate images of real people. The omnibus package also opens trilogue on delaying high-risk AI system deadlines.
OpenAI discontinues its Sora video generation app, API, and website six months after launch, citing unsustainable compute costs. Disney's planned $1 billion Sora partnership also ends, with no investment having closed.
White House Unveils National AI Legislative Framework
The Trump Administration releases a six-pillar national AI legislative framework urging Congress to preempt state AI laws, protect children online, safeguard IP rights, prevent censorship, enable innovation, and develop an AI-ready workforce.
OpenAI releases GPT-5.4 mini and nano, bringing near-flagship performance to smaller, faster models. Mini scores 72.1% on OSWorld-Verified (vs. 75.0% for full GPT-5.4) at $0.75/$4.50 per million tokens, while nano targets high-volume tasks at $0.20/$1.25.
NVIDIA releases Nemotron 3 Super, an open-source 120B-parameter hybrid Mamba-Transformer MoE with 12B active parameters optimized for agentic AI. Delivers 5x higher throughput with a 1M-token context window, open weights, and full training recipes under a permissive license.
The US Senate Sergeant at Arms authorizes ChatGPT, Gemini, and Microsoft Copilot for official use by Senate staff for drafting, research, and analysis. Claude and Grok are notably excluded from the approved list.
OpenAI releases GPT-5.4, unifying the Codex and GPT lines into a single frontier model with a 1M-token context window, native computer use, and conversation compaction for agents. Scores 83.0% on GDPval (up from 70.9%) and 75.0% on OSWorld, surpassing the human baseline.
OpenAI releases GPT-5.3 Instant as the new default model for all ChatGPT users including the free tier, replacing GPT-5.2 Instant. Delivers 26.8% fewer hallucinations on web search queries and reduces unnecessary refusals and defensive preambles.
Perplexity launches Computer, a multi-model agent platform coordinating 19 AI models to autonomously execute complex workflows. Uses Claude Opus 4.6 for orchestration with 400+ app integrations. Available to Perplexity Max subscribers at $200/month.
Google DeepMind releases Gemini 3.1 Pro with a 2x+ reasoning improvement over 3 Pro, scoring 77.1% on ARC-AGI-2. Features dynamic thinking with adjustable depth, a 1M-token context window, and 64K-token output.
Anthropic releases Claude Sonnet 4.6, now the default across free and paid plans. Delivers near-Opus-level performance in coding, computer use, and long-context reasoning with a 1M-token context window at Sonnet-tier pricing ($3/$15 per million tokens).
OpenAI releases GPT-5.3-Codex, combining GPT-5.2-Codex coding performance with GPT-5.2 reasoning at 25% faster speeds. Sets new highs on SWE-Bench Pro and Terminal-Bench. First model rated high for cybersecurity on OpenAI's preparedness framework.
Anthropic releases Claude Opus 4.6, their most capable model yet, featuring a 1M-token context window, agent teams for multi-agent collaboration, conversation compaction, and improved reasoning and coding across benchmarks.
The second International AI Safety Report provides an updated science-based assessment of general-purpose AI risks and safeguards. Notes that companies publishing Frontier AI Safety Frameworks have more than doubled since the 2025 edition.
Mistral AI releases the Mistral 3 family including Mistral Large 3, a 675B-parameter mixture-of-experts model with a 256K context window, fully open-source under Apache 2.0. The model rivals proprietary frontier models on key benchmarks.
Anthropic donates the Model Context Protocol (MCP) to the Agentic AI Foundation under the Linux Foundation. With 97M+ monthly SDK downloads and backing from OpenAI, Google, and Microsoft, MCP has become the universal standard for connecting AI to external tools.
Google DeepMind releases Gemini 3 Pro, their latest flagship multimodal model, with state-of-the-art reasoning, agentic coding capabilities, and availability across AI Studio, Vertex AI, and third-party platforms like Cursor and GitHub.
OpenAI releases GPT-5, a unified system with a built-in reasoning router that automatically selects between a fast model and deeper thinking mode. Hallucinations reduced ~80% vs. o3 in thinking mode. Succeeds GPT-4.5 as the flagship model.
The second wave of EU AI Act obligations takes effect, covering general-purpose AI model requirements, notification obligations, governance structures, and penalties. Full high-risk AI system requirements follow in August 2026.
The OECD and European Commission jointly publish the review draft of their AI Literacy Framework for Primary and Secondary Education, with input from Code.org and educators across 20+ countries. Final version expected first half of 2026.
DeepSeek-R1 Matches Frontier Models at Fraction of Cost
Chinese AI lab DeepSeek releases R1, a 671B-parameter open-source reasoning model developed for under $6M that matches or exceeds OpenAI o1 on math and coding benchmarks, reshaping assumptions about the cost of frontier AI development.