AI Intelligence Report
Top Stories
Anthropic doubles Claude Code’s rate limits after picking up SpaceX’s Memphis supercluster
The headline AI deal of the week: Anthropic has effectively rented Elon Musk’s “Colossus 1” data centre in Memphis, adding more than 300 megawatts of compute โ roughly the equivalent of 220,000 Nvidia GPUs coming online over the next month. The immediate, visible result is that paying Claude Code users are seeing their rate limits doubled. The deeper signal is that the major AI labs are now fighting an arms race for raw electricity and chips, not just talent. For comms purposes, expect “AI’s energy footprint” to escalate as a story over the coming weeks.
OpenAI ships GPT-Realtime-2, plus translation and transcription models โ voice agents finally sound human
OpenAI released three new voice models on May 7. GPT-Realtime-2 is the first voice model with GPT-5โclass reasoning, handling audio input and output in a single loop (rather than transcribing, thinking, then speaking). Two companion models โ GPT-Realtime-Translate (live translation across 70+ input languages and 13 output languages) and GPT-Realtime-Whisper (live transcription) โ round out the launch. The 128K-token context window is four times the previous version, which means longer real conversations are now feasible. This is the model class that powers call-centre and live-translation use cases โ directly relevant for any global comms team coordinating across markets.
Bloomberg: Tencent and Alibaba face slowing growth as AI capex doubles ahead of May 13 earnings
Bloomberg Intelligence is forecasting Tencent’s full-year earnings growth to slow into the low-teen percentage range as AI investment roughly doubles. Tencent reports Q1 results on May 13, with sell-side consensus pointing to revenue near RMB 200.1 billion (up about 11.1% year-on-year) and non-IFRS net profit of about RMB 68.1 billion. The narrative arc to watch is whether the recent share-price rebound can survive a closer look at AI spend versus AI return. Worth pre-positioning talking points on monetisation pathways for Hunyuan, WeChat agents, and cloud.
U.S. government adds Microsoft, Google DeepMind and xAI to its pre-release model testing program
The Center for AI Standards and Innovation (CAISI) signed agreements this week with Microsoft, Google DeepMind and xAI giving the U.S. government the right to evaluate frontier AI models before they ship publicly. This builds on prior CAISI deals with OpenAI and Anthropic from 2024, both of which have now been renegotiated under updated White House directives. National Economic Council director Kevin Hassett separately compared the emerging vetting regime to FDA drug approval. For multinational tech communicators, this is the moment when “voluntary” model testing starts to look mandatory โ and the precedent will travel.
Anthropic commits $200 billion to Google Cloud and chips over five years
Anthropic disclosed a five-year, $200-billion commitment to Google Cloud and Google’s TPU chips. In the same disclosure window, the company said run-rate revenue has reached roughly $30 billion, up from about $9 billion at the end of 2025 โ a tripling in roughly five months. The combination of the SpaceX/Colossus deal (story above) and this Google commitment positions Anthropic as a multi-cloud, multi-chip operator, hedging across Nvidia GPUs, Google TPUs and now also xAI infrastructure.
AI News Roundup
Substack Highlights
Inoreader AI Folder
GPT-Realtime-2 = voice agents finally don’t suck?
The Neuron’s lead story today is GPT-Realtime-2 โ they argue this is the model that makes voice agents production-ready for call centres for the first time. Plus a side note that “Anthropic figured out how to read Claude’s mind,” referring to Anthropic’s Natural Language Autoencoder interpretability work.
OpenAI may not be able to IPO in 2026
David Gerard’s Pivot To AI argues OpenAI’s CFO Sarah Friar has been quietly walking back the 2026 IPO timeline โ the trouble being that going public means publishing your audited financials, and OpenAI’s financials remain “hilariously terrible” by Friar’s own framing. A useful counterweight to the bullish coverage.
Google Chrome force-installs a 4-gigabyte AI model โ and how to get rid of it
Practical, slightly snarky walkthrough of why Chrome is now silently downloading “Gemini Nano” (a 4 GB on-device LLM) for every user, and the steps to opt out. The story is starting to get pickup beyond the AI-skeptic press.
AI just found 15 years’ worth of bugs in Firefox. In weeks!
The Automated newsletter highlights how AI-assisted security tooling โ including Anthropic’s Mythos preview and similar work from Microsoft and Google โ is uncovering long-dormant Firefox vulnerabilities at a pace human researchers couldn’t match. Reinforces the Glasswing narrative: AI cyber capabilities are usable on both offence and defence.
Anthropic adds ‘dreaming’ feature and uncaps agent limits
AI Breakfast reports on two Anthropic moves: an experimental “dreaming” feature (where Claude consolidates and reorganises long-context memories during idle time) and the rate-limit doubling on Claude Code that followed the SpaceX/Colossus deal.
Anthropic just leased Elon’s Memphis supercluster
Not a Bot’s framing of the same Anthropic-Colossus story emphasises the “depth gap” โ they argue OpenAI’s compute lead over Anthropic widens to 3.5x even after this deal. Includes a side note that Wendy’s keeps scaling its AI drive-through deployment while Taco Bell has paused theirs.
Lore Issue #183: Anthropic Taps SpaceX for Massive Compute
Lore Brief’s daily summary leads on the same Anthropic-SpaceX story, plus three sidebars: Google launches “Fitbit Air,” GPT-5.5 Instant reaches all ChatGPT users, and ChatGPT now lives inside Microsoft Excel and Google Sheets natively.
OpenAI wants AI to talk like a human
AI Valley’s analysis of GPT-Realtime-2 frames it as the moment voice AI moves from “annoying script reader” to “credible call-centre operator.” A second story argues a quiet shift is underway in research labs from chatbots toward “world models” โ AI systems that simulate physical environments โ citing recent work at DeepMind and FAIR.
The SpaceXAI Era
An AI Valley think-piece arguing we are entering a phase where Musk’s xAI/SpaceX compute infrastructure is becoming the shared backbone for multiple frontier labs โ Anthropic’s Colossus deal being exhibit A. Side note: the U.S. government wants to inspect frontier models before release.
The Prompt That Makes AI Less Stupid
Bagel Bots shares a prompt template that forces an LLM to critique, rewrite and sharpen its own answer in three explicit passes โ a small workflow trick worth keeping in your prompt library, especially for first drafts of long-form comms copy.
GPT-5.5 remembers (mostly)
Matt Wolfe’s Future Tools roundup highlights GPT-5.5’s improved (but still imperfect) long-term memory, plus a sidebar that Samsung has joined the “trillion-dollar club” off the back of HBM3E memory demand from AI data centres.
Gemini’s theoretical estimates vs actual benchmarks when picking GCE CPU platforms
Kalev Leetaru runs an experiment having Gemini select CPU platforms for Google Compute Engine VMs, then benchmarks them against Gemini’s own performance predictions. Useful as a concrete example of where current LLMs over- or under-estimate cost/performance trade-offs in real infrastructure choices.
Advancing youth safety and wellbeing in EMEA
OpenAI’s “European Youth Safety Blueprint” plus EMEA Youth & Wellbeing Grants โ pre-positioning ahead of expected EU regulatory pressure on AI products used by minors. Useful comparison material if Tencent develops its own youth-safety messaging.
AI Workflows & Tool Watch
Claude Code 2.1.126 fixes a long list of MCP server reliability bugs
Anthropic’s May 1 Claude Code update quietly addressed several issues that were silently breaking MCP servers โ including a memory leak that could push Claude Code’s RAM use past 10 GB when an MCP server wrote non-protocol data to stdout, and a long-running bug where MCP servers connected but failed silently with zero tools. /mcp now shows the tool count for each connected server and flags servers that connected with zero tools. If your MCP setup has been flaky for the last few weeks, this is the release that fixes it.
Perplexity Comet now ships with Claude Sonnet 4.6 by default for Pro, Opus 4.6 for Max
Comet, Perplexity’s agentic browser, now has a model picker. Pro users get Claude Sonnet 4.6 by default; Max users get Claude Opus 4.6, with Gemini 3.1 Pro available as an alternative. iOS pre-orders are open in the App Store. The “Personal Computer on Mac” feature now lets Comet read and edit local files, do voice orchestration, and browse alongside the desktop app โ closer to the agentic-assistant pattern you’ve been tracking.
Reddit MCP server: 36 tools behind Reddit OAuth, three permission tiers
MCPBundles published a hosted Reddit MCP provider that exposes 36+ Reddit API operations as Claude/Claude Code tools, with three explicit permission tiers (read-only, read-write, account-level). For media-monitoring or community-listening workflows, this is the cleanest way to give Claude scoped Reddit access without writing your own scraper. The standalone “reddit-mcp-buddy” project on GitHub is a lighter alternative for one-off subreddit analysis.
n8n + Obsidian: an overnight agent that sorts fleeting notes
An Obsidian community member published a working n8n flow that runs locally overnight, classifies the day’s “fleeting” notes into permanent topic folders using a self-hosted LLM, and drops a daily summary back into Obsidian. The post includes the full n8n blueprint as a downloadable JSON. Worth borrowing the pattern even if you don’t run it overnight โ same structure works for Drafts โ Things 3, or DEVONthink โ Apple Notes.
“Obsidian-as-podcast-feed” via n8n: Have your notes read aloud during commutes
An n8n template now stitches Obsidian โ webhook โ OpenAI text-to-speech โ private podcast feed: every long note you tag becomes an episode you can listen to in your podcast app. Useful for reviewing long internal briefs or PR memos on a flight. The template is one-click installable from the n8n marketplace.
The “self-critique” prompt pattern (via Bagel Bots)
A simple but effective three-pass prompt: (1) write a first draft, (2) critique your own draft against an explicit checklist, (3) rewrite to address the critiques. Bagel Bots packaged it as a copy-paste template; particularly useful for first drafts of statements, FAQs, and internal memos where tone and accuracy both matter.
OpenAI’s safe-Codex playbook: a useful template for governing internal AI agents
OpenAI’s new “Running Codex safely” doc lays out the four pillars they use internally โ sandboxing, approval gates, network egress policies, and “agent-native telemetry” โ to let AI coding agents operate without leaking secrets or running unbounded shell commands. Even if you never deploy Codex, the framework is a clean blueprint for any internal comms about how Tencent should govern AI agents on staff laptops or in production.
Live AI workflows powered by Reddit + Claude MCP โ an ad-ops case study
A case study circulating on r/ClaudeAI documents an agency that compressed weekly Reddit Ads reporting from four hours to thirty minutes by giving Claude direct MCP access to Reddit Ads + their data warehouse, then asking natural-language questions like “why did our CPA spike on Thursday in r/SaaS?” The transferable insight for global comms: the same “Claude + MCP + first-party data” pattern works for press-coverage analysis, crisis tracking, and weekly executive readouts.
For podcast production: GPT-Realtime-Translate at $0.034/minute makes live multilingual feeds plausible
The new GPT-Realtime-Translate model translates speech across 70+ input languages into 13 output languages in real time, at $0.034 per minute (about $2/hour). For your podcast production workflow with Ecamm, this opens up live English-to-Mandarin (or back) translation as an export track โ significantly cheaper than human interpreters and faster than post-production subtitling. Worth a small pilot before locking into next quarter’s content plans.
Tencent Mentions
Bloomberg: AI capex doubling could squeeze Tencent earnings into the low-teens
Bloomberg Intelligence projects full-year earnings growth slowing as AI investment doubles. Earnings drop May 13. Sell-side consensus: revenue โ RMB 200.1B (+11.1% YoY), non-IFRS net profit โ RMB 68.1B (+11.1% YoY).
Cybernews: Hunyuan Hy3-preview launches, smaller and stronger than HY 2.0
The first major model release since the recruitment of Yao Shunyu from OpenAI. Hy3-preview is 295B parameters (down from 400B+ in HY 2.0) but stronger on complex reasoning and code. Notably positioned as “preview” โ full release pending.
Tencent and Alibaba in talks to invest in DeepSeek at a $20B+ valuation
Reported via WSJ and aggregators. DeepSeek’s first outside fundraise. Worth pre-positioning a “why Tencent is investing in adjacent labs as well as building Hunyuan” line for any reactive press queries.
DimSum Daily: Tencent rebound masks AI doubts ahead of May 13 results
Notes the recent share-price recovery is being driven by AI sentiment but warns the May 13 print could puncture the optimism if AI capex commentary lands badly with analysts. Worth a heads-up note to internal IR / corporate comms partners.
SCMP: Tencent pledges new wave of AI investment, betting on WeChat agents
SCMP coverage from earlier in the year continues to circulate as the dominant English-language framing of Tencent’s AI strategy: “AI wave lifts all boats,” with WeChat agents positioned as the consumer flagship. Useful baseline narrative if a reporter pings about Tencent’s agent strategy.
TrendForce: Tencent Cloud raises AI compute, TKE and EMR list prices ~5% effective May 9
Tencent Cloud announced a roughly 5% price hike on AI compute, Tencent Kubernetes Engine, and Elastic MapReduce, effective today (May 9). Joins similar moves by Alibaba, Baidu and Zhipu. Likely to surface in B2B trade press over the weekend; consider a reactive line on cost-pass-through versus margin protection.
36Kr: ByteDance’s hardware push reportedly making Alibaba and Tencent “anxious”
36Kr argues ByteDance is pulling ahead in custom AI silicon and dedicated training infrastructure, putting both Alibaba and Tencent in a defensive posture. The framing is sharp; worth being aware of even if you don’t engage publicly with it.