Fri, Sep 12, 2025

Alibaba Qwen3‑Next‑80B‑A3B launches – 3.7% active params, 10× at 32K; Seedream 4.0 leads

Stay in the loop

Get the Daily AI Primer delivered straight to your inbox. One email per day, unsubscribe anytime.

Executive Summary

  • Alibaba ships Qwen3‑Next‑80B‑A3B: ~3.7% (≈3B) active parameters across 512 experts; prefill 10.6× and decode 10.0× faster at 32K vs Qwen3‑32B; accuracy nears Qwen3‑235B.
  • Runtime/interop – SGLang HiCache reports up to 6× throughput and 80% TTFT cut; vLLM delivers day‑0 optimized kernels for Qwen3‑Next.
  • ByteDance Seedream 4.0 tops Artificial Analysis: ELO 1,205 (editing) and 1,222 (t2i); priced $30 per 1k gens.
  • Google adds gemini‑embedding‑001 to Gemini Batch API – ~50% cheaper; $0.075 per 1M input tokens; async up to 24h; OpenAI SDK compatible.
  • NVIDIA boosts ComfyUI on RTX PCs – up to 40% faster workflows and 3× diffusion; adds Wan 2.2, Qwen‑Image, FLUX.1‑Krea [dev], Hunyuan3D 2.1.
  • OpenAI–Microsoft MOU: nonprofit control; $100B+ equity stake; $50M literacy grants.
  • Ramp AI Index: 44.5% U.S. businesses pay for AI; OpenAI at 36.5%.
  • Why it matters – Evaluate Qwen3‑Next on your tasks; pilot HiCache/vLLM for latency; compare Seedream image quality/cost; schedule Batch embeddings backfills; plan governance amid OpenAI’s restructure.

📑 Table of Contents

🎙️ Voice, Music and Real‑time

A few notable drops: MiniMax Music‑1.5 (4‑min vocals) on FAL/Replicate, EchoX speech‑to‑speech training idea, and ElevenLabs community showcase. Fewer real‑time telephony items today.

MiniMax Music v1.5 goes live on FAL: 4‑minute songs for $0.03

MiniMax (Hailuo) Music 1.5 is now available on FAL with natural‑sounding vocals, multilingual styles, and up to 4‑minute tracks at $0.03 per generation FAL announcement. You can try it in the hosted playground today playground link, with a direct model page for hands‑on testing MiniMax Music v1.5. The release targets fast, low‑cost music creation for apps and content workflows try here.

MiniMax Music 1.5 goes live on FAL: 4‑minute songs for $0.03 with natural vocals

MiniMax (Hailuo) Music v1.5 is now available on FAL with up to 4‑minute tracks, natural vocals, and multilingual/multicultural styles at $0.03 per generation Model launch. You can try it immediately in the hosted playground Playground link, with the model page highlighting quick turnarounds and accessible pricing MiniMax Music v1.5.

ElevenLabs launches community showcase; first 11 get merch + API credits

ElevenLabs kicked off a community showcase and will reward the first 11 approved submissions with exclusive merch and API credits; deadline Sept 17 and winners on Sept 18 How to join. In context of Voice remixing (alpha remixing launch), this gives builders a venue to ship and share production voice work. Submit via the GitHub repo (fork, add project/profile, open PR), and browse live entries to calibrate quality Showcase launch, Showcase gallery, Showcase repo.

EchoX proposes echo training to boost speech‑to‑speech LLM reasoning

The EchoX paper tackles the acoustic‑semantic gap in speech‑to‑speech LLMs by integrating semantic targets during training, aiming to preserve knowledge and reasoning that degrade when operating purely on audio paper page. Trained on roughly 6,000 hours of speech, EchoX reports advanced performance on knowledge QA benchmarks versus text‑only baselines, suggesting a viable path for higher‑fidelity spoken assistants without text detours paper page.

FAL Workflows 2.0: chain image, video and audio models into end‑to‑end pipelines

FAL showcased a complete video made with Workflows 2.0, combining image, video, and audio nodes into a single orchestrated graph—cloneable as a starting point for your own pipelines Workflow demo, Workflow graph, Cinematic workflow. The platform invites builders to remix and extend workflows for richer, multimodal outputs FAL workflows.

ElevenLabs details authentic multilingual voice building on Cloudflare’s AI Avenue

ElevenLabs joined Cloudflare’s AI Avenue to explain how they capture human nuance, emotion, and personality across languages—useful guidance for teams standardizing voice UX in agents and IVR AI Avenue link. Full episode is available for deeper implementation tactics and demos YouTube episode.

FAL Workflows 2.0 shows end‑to‑end cinematic pipeline with Stable Audio

A full video was created entirely inside FAL Workflows, chaining image and video nodes with Stable Audio 2.5 to produce a cinematic output—clone and edit the exact workflow to adapt it for your stack workflow demo, workflow graph, cinematic workflow. In context of Stable Audio, which added fast API and ComfyUI support, this demonstrates practical orchestration of audio with visuals for production‑ready media automation workflows hub.

ElevenLabs launches community showcase with prizes for first 11 projects

ElevenLabs opened submissions for a new community showcase: fork the repo, add your project and author profile, and open a PR; the first 11 approved submissions receive exclusive merch plus API credits how to join. Deadline Sept 17, winners on Sept 18 launch details. Explore live examples to see what’s being built and get inspiration showcase gallery. Source steps and contributing guide are provided to streamline participation GitHub repo.

ElevenLabs on Cloudflare AI Avenue: building authentic multilingual voices

ElevenLabs discussed how they capture human nuance, emotion, and personality across languages, sharing production learnings and pipelines on Cloudflare’s AI Avenue AI Avenue episode. If you want the full session, watch the YouTube recording for details on voice fidelity, cross‑lingual constraints, and deployment considerations YouTube episode, with a quick pointer here as well watch link.


🤖 Embodied AI and Robotics

Sparse but interesting: Physical Intelligence context‑length discussion, Zoox expansion, and Ant Group’s Robbyant R1 humanoid demos. Mostly perspective pieces and product teasers.

Ant Group’s Robbyant R1 humanoid moves toward deployments with 34 DOF and ‘scenario’ bundles

Ant’s wheeled, two‑arm humanoid (110 kg; 1.6–1.75 m; <1.5 m/s; 34 DOF) is shown cooking and giving tours, with claims of mass production and early installs (e.g., a history museum). The go‑to‑market wraps hardware + software + services into pre‑packaged scenarios to ease integration uptime. Ant points to its 300B MoE “Bailing” LLM for step planning and sim‑to‑real training for faster, safer adaptation Robbyant R1 details.

VLA‑Adapter shows tiny‑scale VLA can train in 8 hours on one consumer GPU

A new VLA paradigm uses a lightweight policy with “Bridge Attention” to inject the right vision‑language conditions into action space, avoiding huge pretraining. Results report SOTA with a ~0.5B backbone, fast inference, and training in ~8 hours on a single consumer GPU—lowering the bar for embodied agents Paper summary, Hugging Face paper.

Sergey Levine: fully autonomous home robots are ~5 years out, LLMs provide the ‘common sense’ scaffold

In a wide‑ranging interview, @svlevine argues autonomy at home could arrive in ~5 years, enabled by vision‑language‑action stacks where LLMs inject prior knowledge and task structure. He contrasts tiny robotics context (≈1 s, ~100 ms steps) with human brain efficiency and stresses the deployment data flywheel, simulation, and hardware constraints as the gating factors Podcast overview, Brain vs robots point.


🧮 Chips and Accelerators

Chip chatter centered on inference specialization and new compute visions: Rubin CPX analysis threads and Naveen Rao leaving Databricks to build a low‑cost AI computer.

Rubin CPX cuts memory $/GB >50% with GDDR7; targets million‑token context

NVIDIA’s Rubin CPX swaps HBM for GDDR7 to slash memory cost per GB by more than 50% while maintaining high bandwidth, and is scoped for >1M‑token context windows Bandwidth vs compute, 1M+ window. In context of Rubin CPX (context‑phase GPU debut), this reinforces CPX’s role as a low‑$ memory, high‑throughput accelerator for the attention prefill stage, complementing decode‑optimized parts and enabling cheaper long‑context inference.

OpenAI preps massive compute buys and custom silicon alongside new data centers

OpenAI is locking in compute and chip purchases that could top ~$60B per year, planning ~$18B for a new data center and ~$10B for custom silicon—funding path still evolving with backers WSJ summary. Front‑loaded capex signals a push to bring cost per token down via vertical integration, with implications for GPU demand and future accelerator mix.

Naveen Rao is leaving Databricks to build a next‑generation computer aimed at shrinking AI compute costs and accelerating model progress founder move. For chip buyers, a purpose‑built stack that improves memory locality and $/token economics would pressure GPU incumbents and reshape where training and inference dollars flow.

Alibaba and Baidu pivot to in‑house AI chips; some workloads claim H20 parity

Alibaba (AI accelerator) and Baidu (Kunlun P800) have begun training select models on internal silicon, reportedly matching NVIDIA H20 on some runs, while still relying on NVIDIA for top‑end training in-house chips. This hedges export risk, trims cost, and could erode NVIDIA’s China data‑center share if parity holds at scale.

NVIDIA’s rent‑back chip contracts approach ~$15B through mid‑2025

Multi‑year agreements to rent back NVIDIA’s own GPUs from cloud partners have climbed toward ~$15B by July 2025, up sharply from late 2022 rentback chart. The financing model underlines sustained accelerator scarcity and demand smoothing, but also ties NVIDIA tighter to hyperscaler capacity cycles.

DGX Cloud scales back as external service; refocus on NVIDIA’s internal workloads

NVIDIA is stepping back from positioning DGX Cloud against AWS/Azure; customers cited pricing and channel friction, and the business reportedly plateaued near ~$2B/year. DGX Cloud will mainly support NVIDIA’s own chip design/model work going forward DGX shift. Strategically, this lowers channel conflict while keeping capacity close to R&D.


🗂️ Data, Retrieval and GraphRAG

Data prep and retrieval showed up mainly via GDR data refinement, graph DB shoutouts, and multimodal retrieval. Mostly data cleaning/safety and search foundations; few classic RAG stacks.

DeepMind’s GDR rewrites risky data while preserving style; 0.99 recall on 108 PII types

Generative Data Refinement (GDR) turns messy real data into training‑ready text by rewriting PII/toxicity but keeping natural style and diversity. Reported results: 0.99 recall and 0.80 precision across 108 PII types, with examples that swap real SSNs, API keys, names and detoxify phrasing while leaving safe content untouched paper snippet, method diagram. Full methodology and figures in the paper ArXiv paper. For data teams, this is a practical path to scale high‑quality corpora without blanket filtering that hurts coverage.

MetaCLIP2 lands in Transformers with 329‑language image–text search demos and fine‑tuning

MetaCLIP2 is now usable via Transformers with a working notebook for text→image search and OCR‑aware flows; it supports 329 languages and includes docs plus a demo Space for box‑aware OCR to Markdown conversion notebook link, Transformers docs, Meta CLIP collection. If you need multilingual multimodal retrieval, the examples show end‑to‑end setup and how to ground layout for document search.

Milvus pairs Nano Banana with CLIP to build enterprise‑ready multimodal RAG

Milvus shows how to turn Nano Banana’s image generation into a production multimodal RAG pipeline by embedding text and images with CLIP, storing billions of vectors, and performing fast semantic search to retrieve outfits/props/base assets for generation workflows Milvus blog, Milvus blog post. The walkthrough covers dependency setup, embedding, and retrieval loops to operationalize creative asset search beyond pure prompting.

Firecrawl v2.2.0: 15× faster Map (to 100k URLs), signed webhooks, MCP v3 cloud transport

Firecrawl’s 2.2.0 release focuses on scale and reliability for web data pipelines: Map is 15× faster and supports up to 100k URLs, MCP v3 lands with stable cloud (HTTP transport + SSE), signed webhooks (with failure handling and /extract), per‑API‑key usage tracking, new regions (CA, CZ, IN, IT, PT), and a queue status endpoint release highlights, changelog. If you’re feeding retrieval/indexing stacks, this trims crawl time and tightens event integrity.

R.I.P. naive RAG: HORNET panel makes the case for agentic retrieval

A SHACK15 panel from HORNET.dev argues for moving beyond chunk‑and‑embed RAG to multi‑step, agentic retrieval with a schema‑first API and throughput/recall over single‑turn latency Event page, HORNET site. In context of RAG talks (modern IR talks), this adds concrete patterns for agents that iterate, read whole docs, and operate over scoped or web‑scale data, including on‑prem/VPC deployment and model‑agnostic pipelines for production search.


📊 Evals and Leaderboards

Image editing/text‑to‑image leaderboards, new agent evals, and easy UI eval runners. Notably LiveMCP‑101 shows hard‑task gaps; Weave adds no‑code evals; SimpleQA Verified debuts.

LiveMCP‑101 real‑time agent eval: top models under 60% task success; GPT‑5 scores 39% on hard

A new benchmark, LiveMCP‑101, stress‑tests agents on real app stacks and live tasks; even the best LLMs stay below 60% overall, with GPT‑5 hitting 39.02% on hard tasks Framework intro, Hard task score. The error taxonomy highlights seven common failures: ignoring requirements, overconfident self‑solving, unproductive thinking, wrong tool selection, syntactic errors, semantic errors, and output parsing errors Error taxonomy. Useful for agent gating and regression alerts before production rollouts.

Seedream 4 surges to #2 in Image Edit and #5 in Text‑to‑Image on LMArena

With 43k+ community votes tallied, Seedream 4 lands #2 on the Image Edit leaderboard and #5 on Text‑to‑Image; Gemini 2.5 Flash Image (nano‑banana) remains #1 in both charts Leaderboard update, Text to image rank. You can compare models side‑by‑side and battle them directly in the image arena LMArena image, and Seedream 4 High Res is now available for head‑to‑head tests High res in arena. Real prompts and votes at scale tighten confidence intervals and make rankings more reliable for production picks Leaderboard update.

SimpleQA Verified: Gemini 2.5 Pro at F1 55.6 on parametric factuality (no tools)

Gemini 2.5 Pro records F1 55.6 on SimpleQA Verified, a 1,000‑question benchmark measuring what models know from memory (no tool use, strict answer formatting) Benchmark details. This follows the benchmark’s launch and leaderboard in context of leaderboard debut which established the test’s cleaned labels, balanced topics and abstention‑aware grading. Good for tracking knowledge drift and over‑hedging across releases.

Artificial Analysis: Qwen3‑Next‑80B scores 54 (reasoning) near DeepSeek V3.1; detailed cost tokens

Artificial Analysis places Qwen3‑Next‑80B (Reasoning) at 54 on its Intelligence Index, alongside DeepSeek V3.1 (Reasoning); the non‑reasoning variant scores 45 Model page, AA reasoning page, AA instruct page. Token accounting shows ~100M tokens with reasoning and ~25M without to run the index—slightly less verbose than Qwen3‑235B with reasoning Token usage. Pricing on Alibaba Cloud is noted at $0.5/$6 per 1M in/out tokens for reasoning, and $0.5/$2 without Model page.

Unified Generalization Score leaderboard debuts: Gemini 2.5 Pro leads at 138.8 UGS

A new cross‑arena leaderboard ranks models by a Unified Generalization Score (UGS) spanning Text, Vision and WebDev arenas. Top entries: Gemini 2.5 Pro (138.8), Claude Opus 4.1 Thinking (133.5), GPT‑5‑high (131.6), then OpenAI’s o3 and 4o‑latest UGS leaderboard. The board emphasizes versatility over single‑task peaks—useful for portfolio model selection.

Seedream 4 High Res (4096×4096) added to LMArena image battles

LMArena introduced a 4096×4096 “Seedream 4 High Res” variant so teams can judge outputs at production resolutions in Battle, Side‑by‑Side and Direct modes High res added, Arena links. Check the leaderboard and queue comparisons to see how the high‑res model stacks up versus other top generators Overview leaderboard, LMArena image.

Baidu’s ERNIE‑4.5‑21B‑A3B‑Thinking climbed to #1 on Hugging Face trending models, signaling strong community pull for reasoning‑tuned MoE checkpoints Trending update, Model card. Handy signal for fine‑tune trials and inference bake‑offs when screening open‑weights options.


🧠 Training, RL and Reasoning Methods

Several methods papers and discussion: RL‑driven improvements (AlphaEvolve compute savings), EXIT self‑improve at test time, RL aggregators (AGGLM), diversity via DPP. Mostly reasoning gains.

EXIT trains single‑step policies, then reliably self‑improves over many steps at inference

Meta Superintelligence Labs’ Exploratory Iteration (EXIT) trains only 1‑step updates, then chains 4–16 self‑improvement steps at test time, fixing 170+ math answers in experiments while generalizing to multi‑turn tool use and coding tasks paper first page. The method grows a buffer of informative partial attempts, restarts from them, and scores each update by relative gain (GRPO), improving best‑of‑K without locking training depth paper first page.

Determinantal point processes raise response diversity with minimal single‑try loss

DQO adds a diversity bonus using determinantal point processes on small per‑prompt batches: vectors of candidate answers are spread out, then combined with usual rewards, improving best‑of‑10 results across reasoning, summarization, instruction following, and stories while keeping 1‑sample accuracy largely intact paper summary. Tuning the diversity weight or group size increases variety; pushing too far can nick 1‑shot performance, so the paper recommends moderate settings paper summary.

Pre‑tokenize and localize data to saturate 256‑GPU training

MIT Lincoln Lab shows the bottleneck at 128 nodes (256 GPUs) was input data, not gradient exchange. Two changes—pre‑tokenize and shrink a 2 TB corpus to ~25 GB of token IDs/masks, then copy it to every node—removed network hotspots and delivered near‑linear scaling; data loaders only need to keep one GPU near 100% to realize the gains paper first page paper recap. As models grew, batch size constraints pushed toward model parallelism, but the data path fix alone delivered most of the speedup paper first page.

AlphaEvolve yields measurable training‑time savings for Gemini

Google DeepMind’s Pushmeet Kohli reports AlphaEvolve cut ~0.7% of total compute and sped up Gemini training; the method also showed wins on out‑of‑distribution tests, indicating learned heuristics carry beyond the training distribution compute savings. Small percent savings are large at Gemini scale and compound with other pipeline optimizations compute savings.

Task representations emerge at key tokens and transfer via activation paste

DeepMind finds compact, transient task states appear right before answers; extracting the hidden vector at those tokens and pasting it into a zero‑shot run restores behavior without the full prompt on Gemma V3 4B/12B/27B, with middle layers most useful paper details. Signals are sporadic and local (first chunk of work is strongest); tasks needing long running memory (e.g., counting) don’t compress into a single vector paper details.

Symbolic layer on top of LLMs boosts long‑horizon planning and tool use

CoreThink (“General Symbolics”) wraps LLMs with a symbolic reasoning layer to plan, verify, and tool‑call over long horizons. Reported results: 66.66% on LiveCodeBench v6, 89% on instruction‑following evals, 24.4% on ARC‑AGI‑2, and a companion agentic IDE scoring 62.3% on SWE‑Bench Lite paper link paper details. The approach avoids extra model training costs and targets robustness over many steps by externalizing state and decisions paper details.


🎨 Generative Media and Visual AI

A big week: Seedream 4.0 climbs leaderboards and integrates; Tencent Hunyuan‑Image 2.1 lands in Arena and local quantized; ByteDance HuMo (text+image+audio→video). Mostly image edit/text‑to‑image and unified pipelines.

Seedream 4 jumps to #2 in Image Edit and adds 4K High‑Res in LMArena

With 43k+ votes, Seedream 4 is now #2 on Image Edit and #5 on Text‑to‑Image in LMArena leaderboard update, t2i details; a new “High Res” variant supports 4096×4096 output and is live in head‑to‑head battles high res added, arena link. This builds on its unified edit+generate engine noted earlier initial launch. BytePlus joined a ComfyUI session to walk through the official API node and usage BytePlus livestream, comfy node note. Check rankings and compare models directly in Arena LMArena leaderboard, overview leaderboard.

HunyuanImage 2.1 releases quantized build for 24GB local GPUs

Tencent published an official quantized release of HunyuanImage‑2.1 for local deployment—runs with as little as 24GB VRAM quant release. Model weights and page are up on Hugging Face HF page, HF link, with hosted options also offered via FAL providers for quick trials fal provider. See the model card for capabilities and setup Hugging Face model.

ByteDance unveils HuMo: text+image+audio to high‑fidelity, identity‑preserving video

HuMo is a unified text+image+audio→video system that preserves identity and tightly lip‑syncs speech via a two‑stage recipe (identity injection, then audio cross‑attention with focus‑by‑predicting) method overview, training details. It lets users balance identity vs. lip‑sync vs. motion using time‑adaptive guidance at inference project link. Demos and paper page are live project page.

Gemini 2.5 Flash Image leads community rankings, Photoshop integration coming

Community voting keeps Gemini 2.5 Flash Image (nano‑banana) atop both Image Edit and Text‑to‑Image charts in LMArena snapshots leaderboard update, t2i breakdown, arena filter. A Photoshop integration is slated for this month, signaling deeper creative‑tool reach photoshop note.

Wan2.2‑S2V launches on Replicate for audio‑driven video generation

Replicate added Wan2.2‑S2V, which generates cinematic video from an audio clip plus a reference image—targeting lip‑sync, motion, and camera control in one pipeline replicate listing. Try it here Replicate model.

FAL Workflows ships end‑to‑end cinematic pipelines combining image, video and audio

FAL showcased a full video created entirely with Workflows, chaining image models, video nodes, and Stable Audio in a visual graph workflow demo, builder view, workflows 2.0. You can clone the exact pipeline to remix cinematic workflow or browse models/workflows fal workflows.

Kosmos 2.5 lands in Transformers with OCR‑to‑Markdown and fine‑tuning notebook

Transformers now includes Microsoft’s Kosmos‑2.5 for document understanding; there’s a live demo for OCR with box detection and Markdown export plus a grounded fine‑tuning notebook transformers support, demo + docs. Try the demo and docs here Hugging Face demo, Transformers docs, with a FT walkthrough ready to run fine‑tuning notebook.


💼 Funding and Enterprise Moves

Significant capital and structure updates: OpenAI–Microsoft MOU for PBC, Perplexity raises $200M at $20B, adoption metrics, and Claude memory rollout for teams. Mostly enterprise traction signals.

OpenAI–Microsoft MOU sets up $100B PBC equity stake; nonprofit keeps control

OpenAI and Microsoft signed a non‑binding MOU that would give OpenAI’s nonprofit control at the top while taking roughly a $100B equity stake in a new PBC; it also tees up a $50M grant program for AI literacy and community innovation, with safety oversight remaining with the nonprofit and coordination underway with CA and DE regulators joint details.

OpenAI and NVIDIA plan multibillion‑dollar UK data center build with Nscale

OpenAI and NVIDIA are expected to announce a multi‑billion UK data center investment next week in partnership with Nscale Global Holdings, coinciding with a Trump visit; early reporting urges caution but cites imminent announcements UK report, follow‑up. This scales capacity in context of Oracle capacity OpenAI’s reported $300B multi‑year Oracle pre‑buy. If confirmed, it signals sustained front‑loaded capex for inference and agent workloads.

NVIDIA scales back DGX Cloud to internal focus as revenue plateaus near $2B

NVIDIA is de‑emphasizing DGX Cloud as a public AWS competitor and refocusing it on internal R&D/model work amid pricing pressure and hyperscaler channel tensions; segment revenue hovered near ~$2B by late 2024 and contracts often churned back to main clouds. The 2023 list price was $36,999 per H100/month; AWS later cut H100/A100 pricing by up to 45% strategy report, earnings detail.

McKinsey forecasts $6.7T data center capex by 2030, ~70% AI‑driven

A McKinsey baseline projects ~$6.7T in global data center investment by 2030, with ~70% driven by AI; capex split roughly 60% chips/hardware ($3.1T), 25% power/cooling/network ($1.3T), 15% construction ($0.8T). Inference is expected to dominate workloads by 2030, with staged, modular builds recommended capex summary.

Anthropic rolls out Claude memory to Team/Enterprise with editable summaries

Claude’s new memory for work remembers team processes, clients, and project details; admins can disable it, and users can view/edit summaries and use project‑scoped memory. Import/export flows let teams move memory from ChatGPT or Gemini via Markdown with daily synthesis, preserving work‑relevant items feature rollout, analysis and links, import guide, update note.

NVIDIA’s rent‑back contracts for its own GPUs near $14.5B by mid‑2025

NVIDIA’s multi‑year contracts to rent back its own AI chips from cloud providers (e.g., Microsoft, Oracle, CoreWeave) have grown steadily since late 2022, reaching close to ~$14.5B by July 2025 per NVIDIA/The Information reporting contracts chart.

OpenAI leads U.S. paid enterprise AI adoption at 36.5%

Latest adoption snapshot shows OpenAI leading paid enterprise AI adoption in the U.S. at ~36.5% adoption chart. Complementary web share data has ChatGPT at ~80.9% of AI chatbot site traffic in August 2025, rising since May web share stat.

OpenAI launches Korea office amid strong ChatGPT growth

OpenAI formally launched its Korea presence, citing strong government support and rapid ChatGPT usage growth; leadership highlighted the Seoul kickoff Seoul launch context, OpenAI Korea.

Shopify onboards 335 developers to Hugging Face Enterprise

Hugging Face reports 335 Shopify team members active on its enterprise subscription—an indicator of platform‑level adoption of open‑weights tooling at a 20‑year‑old commerce company HF note.

VS Code Copilot Chat adds BYOK via Hugging Face to run Groq models

GitHub Copilot Chat in VS Code now supports BYOK through Hugging Face Inference Providers, enabling direct access to Groq models with your own key and via HF routing BYOK announcement, VS Code integration, availability note. This reduces lock‑in and moves more enterprise workloads onto open‑weights plus fast inference backends.


🧩 MCP and Agent Interop

MCP moved fast this week: ChatGPT Developer Mode for connectors, security discussions, new transports, and payments in MCP. Mostly connector enablement plus safety concerns.

Calendar‑invite jailbreak exfiltrates private email via ChatGPT MCP tools

A single malicious calendar invite can hijack a ChatGPT session using MCP tools and silently extract private emails—no invite acceptance required. The attacker plants a jailbreak in the calendar description, then when the user asks ChatGPT to review their day, the model reads the booby‑trapped event, follows the injected instructions, and emails out sensitive data. A walkthrough plus an open‑source mitigation shipped the same day exploit thread, EdisonWatch blog, open‑source firewall. This lands in context of Developer Mode (ChatGPT enabled write‑capable MCP client); while approvals are per‑session, “approve, approve, approve” fatigue is real, so expect enterprise policies and sandboxes to tighten.

Vercel launches x402‑mcp: HTTP‑402 payments for MCP tools with sub‑$0.001 minimums

Vercel shipped x402‑mcp, an open protocol that lets agents and MCP servers charge per use via HTTP 402 Payment Required—fees <$0.01 and minimums under $0.001, implemented in ~3 LOC. A full starter template also landed for instant end‑to‑end testing launch thread, Vercel blog post, x402 AI Starter. This makes paid tools first‑class in agent workflows (usage‑metered APIs, paywalled scrapers, premium data calls) without bespoke billing plumbling.

Firecrawl v2.2.0 adds MCP v3 cloud transport and 15× faster Map (100k URLs)

Firecrawl’s 2.2.0 release brings MCP v3 with stable cloud transport (HTTP + SSE), a Map stage that’s 15× faster and scales to 100k URLs, signed webhooks (now with /extract), API‑key usage tracking, new regions (CA, CZ, IN, IT, PT, more) and a queue status endpoint release thread, changelog. For agent stacks, that means more reliable MCP connectivity, better observability, and lower latency pipelines for crawl→extract→summarize loops.

Manus brings Hugging Face MCP into agents, exposing 2M models and 500k datasets

Manus integrated Hugging Face via MCP so agents can call into 2M+ models, ~500k datasets, ~500k apps and ~100k papers directly from tool calls—broadening in‑agent retrieval and inference without hardcoding SDKs HF MCP in Manus. This is a big interop win: one connector unlocks the HF ecosystem for planning, evals, and hybrid open/hosted runs.

Genspark AI Browser launches with MCP Store (700+ tools) and 169 on‑device models

Genspark’s agentic browser ships an MCP Store that plugs into 700+ tools (Discord, GitHub, Notion, Slack, etc.) and lets users run 169 open‑weight models locally for private, offline assist. It adds an Autopilot mode for autonomous browsing plus shopping agents that compare products and deals agentic browser, download page, launch article. Useful where data residency or no‑cloud constraints block standard connectors.

ChatGPT ‘API tool’ hint reveals Developer Mode toggle for unverified MCP connectors

A UI hint labelled “API tool” in the ChatGPT web app maps to Developer Mode that enables unverified MCP connectors—useful for internal tooling and lab trials before store approval. Screenshot shows the hint active while targeting GPT‑5 in chat UI screenshot. Pair this with stricter approvals and scoped permissions if you plan to expose write‑capable tools.


⚙️ Serving and Inference Systems

Strong emphasis on long‑context serving: SGLang HiCache, vLLM/SGLang day‑0 for Qwen3‑Next, determinism posts, and throughput/TTFT wins. Mostly infra‑level speedups and cache hierarchies.

XQuant rematerializes KV to slash LLM decode memory by up to 12.5× with near‑FP16 quality

A UC Berkeley–led paper proposes caching low‑bit input activations X and rematerializing Keys/Values on the fly, reporting 10× memory savings at ~0.01 perplexity drop and 12.5× at ~0.1 vs FP16; XQuant‑CL further compresses by storing tiny cross‑layer deltas, trading cheap compute for fewer memory reads (decode is memory‑bound) overview thread, chart explainer, KV vs XQuant figure. This directly targets the KV‑cache bottleneck that grows with context length and batch, improving long‑context serving readiness. Full details in the arXiv preprint ArXiv paper.

vLLM + PyTorch adopt disaggregated prefill/decoding to boost large‑scale inference efficiency

Meta’s vLLM disaggregated implementation highlights efficiency gains by separating high‑bandwidth prefill from latency‑sensitive decode at scale PyTorch note, echoing the prefill‑vs‑bandwidth insight that prefill "pounds matrices" and is limited by memory traffic prefill explainer. This builds on prior best practices to split pools and scheduling for long‑context serving, in context of GKE split (prefill/decode pool recipe). Community infra leads note they’ve championed this design pattern since 2023 @haoailab comment. The result: better hardware utilization, lower tail latency for decode, and improved throughput under long prompts without overprovisioning decode kernels.

Qwen3‑Next hybrid architecture forces deep inference‑engine updates in SGLang and vLLM

Adapting Alibaba’s hybrid Qwen3‑Next‑80B‑A3B required substantial inference‑engine work: ~6,000 lines in SGLang (adding Mamba‑style support) and ~2,500 lines in vLLM despite prior Mamba infra Zhihu breakdown. Analysts note these hybrid designs can be much faster in long‑context scenarios but demand operator‑ and framework‑level optimizations to fully realize gains Zhihu breakdown. Model context and deployment notes: native 256k window, FP8 fits on a single H200, and observed token usage profiles for reasoning vs non‑reasoning variants during arena runs arena metrics, model pages. This wave signals that upcoming hybrid/MoE/linear‑attention stacks will land with day‑0 engine changes rather than drop‑in kernels.


🛡️ Security, Safety and Governance

High‑signal MCP exploit demos, FTC inquiries, and updated Model Spec guidance (agents’ autonomy bounds, safe completions). Focus is on risk posture for agentized systems.

0‑click calendar invite jailbreak exfiltrates email via ChatGPT MCP

Attackers showed they can exfiltrate private email using only a victim’s address by planting a jailbreak in a calendar invite—no acceptance needed—then having ChatGPT’s MCP tools read the invite and execute the attacker’s instructions (approvals fatigue risk). This lands right after ChatGPT opened unverified connectors in Developer Mode, in context of Dev Mode MCP. Demo and write‑up show the steps and an open‑source mitigations repo exploit demo, security blog post, GitHub repo.

FTC opens inquiry into OpenAI, Meta, Alphabet, Snap, Character AI and xAI

The U.S. FTC has launched an inquiry into leading chatbot providers, probing harm testing practices, handling of user inputs, and engagement monetization—amid reports of safety incidents like inappropriate chats with minors FTC inquiry.

OpenAI updates Model Spec with Root authority, agent autonomy bounds and Safe Completions

OpenAI revised its Model Spec: Root now outranks System (non‑overridable core principles), adds rules for agent autonomy (act only within agreed scope; manage/communicate side effects), introduces No Other Objectives, and shifts refusals toward Safe Completions. Details and examples are in the release notes and spec snapshot release notes summary, with links to the updated docs Model release notes and the full spec page OpenAI Model Spec.

OpenAI details joint security testing with US CAISI and UK AISI; two ChatGPT Agent flaws found

OpenAI outlined its collaboration with US CAISI and UK AISI on secure deployment: red‑team style testing, end‑to‑end product security evals, and rapid feedback loops surfaced two novel vulnerabilities in ChatGPT Agent (initially assessed as not easily exploitable) that were mitigated post‑discovery OpenAI update, OpenAI blog post.

Parental “Vessel controls” spotted in ChatGPT hint at account‑level safety management

Hidden settings surfaced for “Vessel/Amphora Controls” show invite‑based owner/member roles with toggles for safe model policy, content restrictions, time limits, and access controls (memory, voice, model training). This points to forthcoming parental/guardian or managed‑account safety features settings screenshot.

Vessel controls modal

US HHS urges all employees to use ChatGPT, with HIPAA limits

HHS directed agency‑wide adoption of ChatGPT to boost productivity—one of the broadest federal deployments to date—while reminding units under HIPAA not to disclose protected health information to the tool policy report.


🏗️ Cloud, Capex and Capacity

Big infra moves: OpenAI+NVIDIA eye UK datacenter investments with Nscale; Oracle AI backlog; McKinsey’s multi‑trillion DC forecast; China incumbents train on in‑house chips. Heavy capacity news.

OpenAI lines up ~$60B/yr chip buys, adds $18B data‑center and $10B ASIC plans

OpenAI is preparing to pre‑commit massive capacity: ~$60B/year for GPUs, ~$18B for a new data center, and $10B for custom silicon, even as revenue ($13B) lags and the firm targets $100B by 2028 and $200B by 2030 WSJ analysis. The scale underscores an aggressive build‑now, monetize‑later stance in context of Oracle pact five‑year compute prebuy, with recent backers committing ~$50B and another ~$19B contingent on a structural shift WSJ analysis.

McKinsey: $6.7T data‑center capex by 2030, ~70% AI‑driven

McKinsey pegs cumulative global data‑center investment at $6.7T by 2030, with ~70% tied to AI workloads McKinsey forecast. Baseline assumes ~156 GW of AI capacity by 2030 (125 GW added in 2025–2030). Capex mix: 60% chips/hardware ($3.1T), 25% power/cooling/network ($1.3T), 15% construction ($0.8T). Inference expected to dominate AI work by 2030 McKinsey forecast.

UK buildout with Nscale slots into rising AI‑factory wave

The expected OpenAI/NVIDIA + Nscale announcement signals another ‘AI factory’ node in Europe’s power‑constrained landscape UK data center plan, Nscale tie‑up. Coupled with hyperscaler resale and rent‑back dynamics Contract growth chart and OEM pivots like DGX Cloud’s refocus DGX Cloud scale‑back, the pattern points to diversified procurement and site control to secure long‑term AI capacity.

Alibaba and Baidu start training models on in‑house chips, trimming Nvidia reliance

Alibaba (own AI accelerator) and Baidu (Kunlun P800) have begun training select models on their chips, with internal chatter that performance now matches Nvidia’s H20 for some workloads In‑house chips shift. The move, driven by export curbs and self‑sufficiency goals, still leaves frontier runs on Nvidia but signals medium‑term pressure on H20 demand in China Alibaba/Baidu update.

OpenAI and NVIDIA to back multi‑billion UK datacenters with Nscale

OpenAI and NVIDIA are expected to announce multi‑billion dollar UK data center investments with London‑based Nscale Global Holdings next week, coinciding with a high‑profile US visit UK data center plan, Nscale tie‑up. The push signals fresh AI capacity build‑out in the UK, with scale and partners still to be detailed.

Nvidia scales back DGX Cloud as costs and channel tensions bite

Nvidia is pulling back its DGX Cloud ambitions to avoid competing with key cloud partners and due to pricing headwinds; the service will focus on Nvidia’s internal research and model work DGX Cloud scale‑back, Q2 FY26 note. Earlier list‑price economics ($36,999/H100‑month) were tolerable under scarcity but less so after public cloud price cuts, and multi‑billion ‘cloud commitments’ are no longer grouped under DGX Cloud in reporting Q2 FY26 note.

Nvidia’s rent‑back contracts for its AI chips approach $15B by mid‑2025

Nvidia’s multiyear contracts to rent back its own AI GPUs from hyperscalers (e.g., Microsoft, Oracle, CoreWeave) climbed steadily from late 2022 to nearly $15B by mid‑2025, underscoring demand for capacity even via re‑renting Contract growth chart.

Hangzhou emerges as an AI capacity hub with Alibaba’s $52B and fresh regional funds

Hangzhou’s AI build‑out accelerates: Alibaba pledges $52B for AI/cloud over 3 years; Zhejiang’s 2024 science spend hit ~$12B; new innovation funds total ~$28B—pulling talent post‑DeepSeek and lowering costs versus Beijing/Shanghai Hangzhou AI hub. Regional capital and policy tailwinds translate into local compute, labs, and accelerator ecosystems.


🧰 Agents and Coding Tooling

Lots of practical agent/coder updates: Replit Agent v3, Cursor Tab model, Claude Code SDK hooks/tools, Cline, Zed integrations. Focus on agentic planning, subagents, and IDE workflows.

Claude Code SDK adds custom tools, permissions hooks and a full docs refresh

Anthropic shipped in‑process custom tools (as MCP) and hooks like PreToolUse/PostToolUse for permissions, logging and bespoke control, plus 10 new guides and updated references, in context of LangGraph 80% prior eval wins SDK update, Custom tools docs, and detailed permissions guidance in Permissions guide. Highlights: define a function → register as an MCP tool → callable by Claude Code; add gatekeeping or audits with hooks; and use new examples across TS/Python. See the refreshed overview and guides in Overview - Anthropic and hook patterns in Handling Permissions.

Anthropic’s “tools for agents” guide lifts Slack accuracy 67→80% and Asana 80→86%

A practical masterclass shows how to prototype MCP servers, evaluate with realistic workflows, then let Claude Code refactor tools based on transcripts—yielding Slack tools up from 67%→80% and Asana 80%→86% via clearer schemas/descriptions Guide heads‑up, Key takeaways, Evaluation loop. The post stresses concise namespaces, human‑readable returns, token‑efficient paging, and LLM‑as‑judge verification. Full write‑up: Anthropic blog.

Qwen Code v0.0.10/11: subagents, TODO tool, Welcome‑Back summaries, and stability fixes

Alibaba’s Qwen Code drops two rapid releases that add subagents for task decomposition, a Todo Write tool, a “Welcome Back” project summary on reopen, terminal bench stress testing, fewer agent loops/retries, better MCP/OAuth, and improved memory/session management Release highlights. The "Welcome Back" dashboard shows plan progress with resumable tasks Welcome back. Task runs display timing and token stats (e.g., 16.9s, 15,988 tokens) and tool success rates, aiding debugging Release highlights. The Todo writer is now a first‑class utility for tracking work items Todo tool.

Cursor Tab ships fewer, smarter completions; background agents land in GitHub workflows

Cursor’s Tab model is “pure magic,” emphasizing fewer but smarter completions trained on live production data Cursor Tab. Teams also showcased background agents running from GitHub comments and Actions, including a composable action to run Cursor/Claude/Gemini/Codex prompts on PRs Comment agent demo, Event recap, with the open‑source action here GitHub repo. These patterns favor async, CI‑native agent operations over chat‑style interaction.

Cline v3.28 improves GPT‑5 coding reliability and extends free Grok access

Cline tuned prompts for GPT‑5 so multi‑step coding tasks stay on track (better context retention, fewer derailments), and extended free Grok access. It also adds ESC to cancel operations, real‑time task history sync, Windows/PowerShell Deep Planning, and fixes for Anthropic extended thinking, LiteLLM caching, and Gemini rate limiting Release heads‑up, Changelog, GPT‑5 tuning note. Full notes in Cline blog and CHANGELOG.

Vercel launches x402‑mcp: open payments for agents with sub‑$0.001 minimums

x402‑mcp brings HTTP 402 “Payment Required” semantics to agents/MCP servers via the AI SDK, advertising <$0.01 fees and sub‑$0.001 minimums. It’s account‑less/anonymous and takes ~3 LOC to implement, plus a full‑stack starter ships alongside Protocol launch, Starter link. Details in Vercel blog and the template at x402 AI Starter.

Zed adds Plan Mode for Claude Code, SSH for third‑party agents, and Gemini CLI fixes

Zed shipped a Preview with Plan Mode and other Claude Code modes, SSH support for third‑party agents, Gemini ACP server proxying via Zed config, plus keybindings and bug fixes Zed Preview. Note: users on 0.203–0.204 may need a one‑time manual update due to an auto‑updater issue that’s already been patched Updater note.

VS Code Copilot Chat adds BYOK for Groq models via Hugging Face providers

Developers can now “bring your own key” to access Groq models directly inside VS Code’s Copilot Chat using Hugging Face inference providers, widening model choice in the IDE Groq in VS Code, BYOK note, Everywhere push.

Firecrawl v2.2.0: MCP v3 stable cloud, 15× faster Map (to 100k URLs), signed webhooks

Agent builders get faster, safer crawling: MCP v3 with stable cloud transport, Map 15× faster up to 100k URLs, signed webhooks (+/extract), usage tracking by API key, more regions, and a queue status endpoint Release thread. Full changelog in Firecrawl.

Codex Launcher brings OpenAI Codex CLI into JetBrains IDEs with auto/full‑auto modes

An unofficial JetBrains plugin shells to the OpenAI Codex CLI, adding model/effort selection, file handling modes, notifications, and a "Full Auto" option; requires installing Codex CLI separately Plugin overview. Plugin page: Codex Launcher.


🚀 New and Updated Models

Heavy on Qwen3‑Next (80B ultra‑sparse MoE), plus sightings of GPT‑5‑high‑new and multiple open models (SpikingBrain1.0, Hunyuan‑MT). Mostly LLMs and translation; vision/music went to their own sections.

GPT‑5‑high‑new spotted; API rate limits massively increased

Developers surfaced a "gpt‑5‑high‑new" label in Codex/Cursor UIs, described as relying on built‑in reasoning defaults; it’s visible but not generally usable yet CLI spot codex release GitHub release. In parallel, OpenAI increased GPT‑5 and GPT‑5‑mini API limits dramatically—Tier 1 jumps from 30K to 500K TPM (1.5M batch), with higher tiers up to 4M TPM, signaling readiness for heavier production traffic rate limits rate limit note. Several tools report better long‑session coding stability after prompt tuning for GPT‑5, consistent with a reasoning‑forward profile cline update.

Qwen3‑Next 80B adds pricing, evals and long‑context details

New today: Qwen3‑Next 80B (A3B) scores 54 on the Artificial Analysis Intelligence Index, within striking distance of top closed models, with a native 256k context window and only ~3.8% of parameters active per token model overview eval details. Pricing on Alibaba Cloud lands at $0.5 per 1M input tokens and $6 per 1M output for the reasoning variant (non‑reasoning $0.5/$2), undercutting Qwen3‑235B by ≥25% depending on workload model overview. Analysts also highlight strong working memory and better multi‑turn stability, with trade‑offs in instruction following and hallucination under heavy token budgets Zhihu analysis infra complexity. This builds on the initial launch that introduced the ultra‑sparse MoE architecture and 10× throughput gains; today’s updates quantify performance, cost, and 256k context, and note the FP8 footprint fits on a single H200 eval details.

Meta’s MobileLLM‑R1 (950M) brings on‑device reasoning gains

MobileLLM‑R1‑950M, a sub‑1B "edge reasoning" model, posts sizable small‑model gains: ~5× higher MATH accuracy vs Olmo‑1.24B and ~2× vs SmolLM2‑1.7B, while using only 4.2T pretraining tokens (~12% of Qwen3’s 36T) model overview. It’s available on Hugging Face under a FAIR non‑commercial research license Hugging Face model, with community demos showing quick chat apps and integration into lightweight UIs chat demo.

Kimi K2‑0905 update posts 69.2% on SWE‑bench Verified

Moonshot’s Kimi K2‑0905 (1T‑param MoE, 256K context) reports 69.2% on SWE‑bench Verified, with solid showings on SWE‑Dev (66.6%) and notable gains on Terminal‑Bench; tool‑calling and front‑end coding were highlighted as improved benchmarks chart. The model remains competitive with top proprietary systems on several engineering benchmarks while offering an open weights–friendly posture for evaluation.

Tencent’s Hunyuan‑MT + Chimera tops WMT2025 with fusion aggregator

Tencent unveiled Hunyuan‑MT‑7B and a companion Chimera‑7B that fuses multiple candidate translations at inference to produce a stronger output, ranking 1st on 30 of 31 WMT2025 language pairs tech report. The pipeline mixes broad pretraining, MT‑focused curation, SFT, and a ‘weak‑to‑strong’ RL stage where Chimera learns to select/merge hypotheses—yielding higher quality under the same model size. The release targets bidirectional translation across 33 languages and emphasizes robustness on low‑resource pairs tech report.

On this page

Executive Summary
🎙️ Voice, Music and Real‑time
MiniMax Music v1.5 goes live on FAL: 4‑minute songs for $0.03
MiniMax Music 1.5 goes live on FAL: 4‑minute songs for $0.03 with natural vocals
ElevenLabs launches community showcase; first 11 get merch + API credits
EchoX proposes echo training to boost speech‑to‑speech LLM reasoning
FAL Workflows 2.0: chain image, video and audio models into end‑to‑end pipelines
ElevenLabs details authentic multilingual voice building on Cloudflare’s AI Avenue
FAL Workflows 2.0 shows end‑to‑end cinematic pipeline with Stable Audio
ElevenLabs launches community showcase with prizes for first 11 projects
ElevenLabs on Cloudflare AI Avenue: building authentic multilingual voices
🤖 Embodied AI and Robotics
Ant Group’s Robbyant R1 humanoid moves toward deployments with 34 DOF and ‘scenario’ bundles
VLA‑Adapter shows tiny‑scale VLA can train in 8 hours on one consumer GPU
Sergey Levine: fully autonomous home robots are ~5 years out, LLMs provide the ‘common sense’ scaffold
🧮 Chips and Accelerators
Rubin CPX cuts memory $/GB >50% with GDDR7; targets million‑token context
OpenAI preps massive compute buys and custom silicon alongside new data centers
Naveen Rao exits Databricks to found a compute company focused on cheaper AI
Alibaba and Baidu pivot to in‑house AI chips; some workloads claim H20 parity
NVIDIA’s rent‑back chip contracts approach ~$15B through mid‑2025
DGX Cloud scales back as external service; refocus on NVIDIA’s internal workloads
🗂️ Data, Retrieval and GraphRAG
DeepMind’s GDR rewrites risky data while preserving style; 0.99 recall on 108 PII types
MetaCLIP2 lands in Transformers with 329‑language image–text search demos and fine‑tuning
Milvus pairs Nano Banana with CLIP to build enterprise‑ready multimodal RAG
Firecrawl v2.2.0: 15× faster Map (to 100k URLs), signed webhooks, MCP v3 cloud transport
R.I.P. naive RAG: HORNET panel makes the case for agentic retrieval
📊 Evals and Leaderboards
LiveMCP‑101 real‑time agent eval: top models under 60% task success; GPT‑5 scores 39% on hard
Seedream 4 surges to #2 in Image Edit and #5 in Text‑to‑Image on LMArena
SimpleQA Verified: Gemini 2.5 Pro at F1 55.6 on parametric factuality (no tools)
Artificial Analysis: Qwen3‑Next‑80B scores 54 (reasoning) near DeepSeek V3.1; detailed cost tokens
Unified Generalization Score leaderboard debuts: Gemini 2.5 Pro leads at 138.8 UGS
Seedream 4 High Res (4096×4096) added to LMArena image battles
Hugging Face trending: ERNIE‑4.5‑21B‑A3B‑Thinking hits #1
🧠 Training, RL and Reasoning Methods
EXIT trains single‑step policies, then reliably self‑improves over many steps at inference
Determinantal point processes raise response diversity with minimal single‑try loss
Pre‑tokenize and localize data to saturate 256‑GPU training
AlphaEvolve yields measurable training‑time savings for Gemini
Task representations emerge at key tokens and transfer via activation paste
Symbolic layer on top of LLMs boosts long‑horizon planning and tool use
🎨 Generative Media and Visual AI
Seedream 4 jumps to #2 in Image Edit and adds 4K High‑Res in LMArena
HunyuanImage 2.1 releases quantized build for 24GB local GPUs
ByteDance unveils HuMo: text+image+audio to high‑fidelity, identity‑preserving video
Gemini 2.5 Flash Image leads community rankings, Photoshop integration coming
Wan2.2‑S2V launches on Replicate for audio‑driven video generation
FAL Workflows ships end‑to‑end cinematic pipelines combining image, video and audio
Kosmos 2.5 lands in Transformers with OCR‑to‑Markdown and fine‑tuning notebook
💼 Funding and Enterprise Moves
OpenAI–Microsoft MOU sets up $100B PBC equity stake; nonprofit keeps control
OpenAI and NVIDIA plan multibillion‑dollar UK data center build with Nscale
NVIDIA scales back DGX Cloud to internal focus as revenue plateaus near $2B
McKinsey forecasts $6.7T data center capex by 2030, ~70% AI‑driven
Anthropic rolls out Claude memory to Team/Enterprise with editable summaries
NVIDIA’s rent‑back contracts for its own GPUs near $14.5B by mid‑2025
OpenAI leads U.S. paid enterprise AI adoption at 36.5%
OpenAI launches Korea office amid strong ChatGPT growth
Shopify onboards 335 developers to Hugging Face Enterprise
VS Code Copilot Chat adds BYOK via Hugging Face to run Groq models
🧩 MCP and Agent Interop
Calendar‑invite jailbreak exfiltrates private email via ChatGPT MCP tools
Vercel launches x402‑mcp: HTTP‑402 payments for MCP tools with sub‑$0.001 minimums
Firecrawl v2.2.0 adds MCP v3 cloud transport and 15× faster Map (100k URLs)
Manus brings Hugging Face MCP into agents, exposing 2M models and 500k datasets
Genspark AI Browser launches with MCP Store (700+ tools) and 169 on‑device models
ChatGPT ‘API tool’ hint reveals Developer Mode toggle for unverified MCP connectors
⚙️ Serving and Inference Systems
XQuant rematerializes KV to slash LLM decode memory by up to 12.5× with near‑FP16 quality
vLLM + PyTorch adopt disaggregated prefill/decoding to boost large‑scale inference efficiency
Qwen3‑Next hybrid architecture forces deep inference‑engine updates in SGLang and vLLM
🛡️ Security, Safety and Governance
0‑click calendar invite jailbreak exfiltrates email via ChatGPT MCP
FTC opens inquiry into OpenAI, Meta, Alphabet, Snap, Character AI and xAI
OpenAI updates Model Spec with Root authority, agent autonomy bounds and Safe Completions
OpenAI details joint security testing with US CAISI and UK AISI; two ChatGPT Agent flaws found
Parental “Vessel controls” spotted in ChatGPT hint at account‑level safety management
US HHS urges all employees to use ChatGPT, with HIPAA limits
🏗️ Cloud, Capex and Capacity
OpenAI lines up ~$60B/yr chip buys, adds $18B data‑center and $10B ASIC plans
McKinsey: $6.7T data‑center capex by 2030, ~70% AI‑driven
UK buildout with Nscale slots into rising AI‑factory wave
Alibaba and Baidu start training models on in‑house chips, trimming Nvidia reliance
OpenAI and NVIDIA to back multi‑billion UK datacenters with Nscale
Nvidia scales back DGX Cloud as costs and channel tensions bite
Nvidia’s rent‑back contracts for its AI chips approach $15B by mid‑2025
Hangzhou emerges as an AI capacity hub with Alibaba’s $52B and fresh regional funds
🧰 Agents and Coding Tooling
Claude Code SDK adds custom tools, permissions hooks and a full docs refresh
Anthropic’s “tools for agents” guide lifts Slack accuracy 67→80% and Asana 80→86%
Qwen Code v0.0.10/11: subagents, TODO tool, Welcome‑Back summaries, and stability fixes
Cursor Tab ships fewer, smarter completions; background agents land in GitHub workflows
Cline v3.28 improves GPT‑5 coding reliability and extends free Grok access
Vercel launches x402‑mcp: open payments for agents with sub‑$0.001 minimums
Zed adds Plan Mode for Claude Code, SSH for third‑party agents, and Gemini CLI fixes
VS Code Copilot Chat adds BYOK for Groq models via Hugging Face providers
Firecrawl v2.2.0: MCP v3 stable cloud, 15× faster Map (to 100k URLs), signed webhooks
Codex Launcher brings OpenAI Codex CLI into JetBrains IDEs with auto/full‑auto modes
🚀 New and Updated Models
GPT‑5‑high‑new spotted; API rate limits massively increased
Qwen3‑Next 80B adds pricing, evals and long‑context details
Meta’s MobileLLM‑R1 (950M) brings on‑device reasoning gains
Kimi K2‑0905 update posts 69.2% on SWE‑bench Verified
Tencent’s Hunyuan‑MT + Chimera tops WMT2025 with fusion aggregator