← Back to feed
AI

Google ships Gemini 3.1 Ultra — 2M tokens, native multimodal, sandboxed code

Google's marquee release of 2026: a 2M-token context window that ingests text, image, audio and video in a single forward pass — no stitched pipelines. Sundar Pichai demoed a sandboxed Code Execution tool that writes, runs and tests Python mid-conversation. On MMMU and VideoMME, Ultra outpaces GPT-5.4; on LM Arena it briefly hit #1 before GPT-5.4 reclaimed top. Available day-one in AI Studio and Vertex, with a 200K 'Flash' tier free up to 1M requests/day.

GoogleGeminiMultimodalLong ContextLM Arena

Why it matters

2M-token native multimodal with sandboxed code execution is the configuration that turns Gemini into a real alternative to GPT-5.4 for agentic workflows — not a catch-up release. Developer tooling built on Gemini should see genuine differentiation from here, especially for video/audio-heavy use cases. Google's distribution advantages (Workspace, Android, Search) now have a model worth distributing.

Impact scorecard

8.8/10
Stakes
8.5
Novelty
8.5
Authority
9.0
Coverage
9.5
Concreteness
9.0
Social
9.0
FUD risk
2.0
Coverage58 outlets · 12 tier-1
The Verge, TechCrunch, Ars Technica, Wired, CNBC, Bloomberg, …
X / Twitter31,000 mentions
@sundarpichai · 42,000 likes
@JeffDean · 18,000 likes
Reddit5,800 upvotes
r/MachineLearning
r/Bard, r/MachineLearning, r/singularity

Trust check

high

Google primary announcement + day-1 independent benchmarks from Artificial Analysis and LM Arena. MMMU/VideoMME numbers reproducible via public API.

Primary source ↗