Google ships Gemini 3.1 Flash TTS — 70+ languages, Elo 1211 on Artificial Analysis leaderboard
·Google Blog
Google rolled out Gemini 3.1 Flash TTS starting April 15 across Gemini API, AI Studio, Vertex AI, and Google Workspace (Google Vids). It supports more than 70 languages, natural-language 'audio tags' for controlling vocal style, pace and delivery mid-sentence, native multi-speaker dialogue, and scene direction. Every generated clip is watermarked with SynthID. On the Artificial Analysis TTS leaderboard, Flash TTS landed an Elo score of 1211, placing it in the 'most attractive quadrant' for quality-per-dollar and directly challenging ElevenLabs' pricing premium. Google did not publish exact latency or pricing numbers.
googledeepmindgeminittselevenlabs
Why it matters
TTS has been ElevenLabs' moat for 18 months. An Elo of 1211 on Artificial Analysis plus 70-language coverage bundled into the Gemini API at Google's usual price aggression threatens that moat directly. For any product that ships conversational agents — support, accessibility, localization, content narration — the default TTS vendor conversation gets rewritten this quarter. Expect ElevenLabs to either cut API pricing by 40%+ or pivot toward voice-cloning/creator tooling.
Impact scorecard
7.6/10
Stakes
7.5
Novelty
7.5
Authority
8.5
Coverage
7.0
Concreteness
8.0
Social
7.0
FUD risk
2.0
Coverage22 outlets · 3 tier-1
The Verge, TechCrunch, Ars Technica, VentureBeat
X / Twitter6,000 mentions @GoogleDeepMind · 8,000 likes
Reddit900 upvotes r/MachineLearning
r/MachineLearning, r/singularity
Trust check
high
Primary-source Google announcement; Artificial Analysis leaderboard is independent and verifiable. Elo 1211 is corroborated live. No FUD flags.
Converge Bio closed a $25M oversubscribed Series A led by Bessemer Venture Partners with TLV Partners, Saras Capital and Vintage Investment Partners participating; execs from Meta, OpenAI and Wiz joined as individual LPs. The company builds generative models trained on DNA, RNA and protein sequence data, with three commercial systems: antibody design, protein yield optimization, and biomarker/target discovery. Traction: 40+ programs with over a dozen pharma/biotech customers across the US, Canada, Europe and Israel, expanding into Asia. Case studies include a 4-to-4.5x protein-yield boost in a single computational pass, and antibodies with single-nanomolar binding affinity. Headcount grew from 9 in Nov 2024 to 34 today. Prior seed was $5.5M in 2024. CEO Dov Gertz founded the company.
NVIDIA released Nemotron OCR v2 on April 17 — an 84M-parameter unified multilingual OCR model trained primarily on 12.2 million synthetic images generated via a modified SynthDoG pipeline, plus ~680K real-world scans. It handles English, Simplified and Traditional Chinese, Japanese, Korean and Russian in a single model (no language detection needed). On OmniDocBench it processes 34.7 pages per second on a single A100 — 28x faster than PaddleOCR v5's server mode at 1.2 pages/s — while holding competitive normalized-edit-distance accuracy. On the SynthDoG multilingual benchmark it dominates: 0.046 NED in Japanese vs v1's 0.723, 0.047 in Korean vs v1's 0.923. Weights and the training dataset are public under NVIDIA Open Model License and CC-BY-4.0.
METR and Epoch AI released MirrorCode, a benchmark that tests whether AI can autonomously reimplement complex real-world software from specification. The headline result: Claude Opus 4.6 successfully reimplemented gotree — a bioinformatics toolkit with roughly 16000 lines of Go and 40+ commands — an effort estimated to take a human engineer 2 to 17 weeks. The benchmark spans 20+ programs across Unix utilities, cryptography and compression. The release also previews a Google DeepMind taxonomy of six attack genres on AI agents (content injection, semantic manipulation, cognitive state, behavioral control, systemic, human-in-the-loop) and Ryan Greenblatt's revised estimate that full AI R&D automation by end-2028 now has 30% probability, up from 15%, citing verifiable-software-task self-improvement loops.