Fresh stories
Claude Opus 4.7 releases with xhigh effort, /ultrareview, and 3x vision resolution
Claude Opus 4.7 is now generally available across Claude, the API, and major clouds with xhigh effort, higher-resolution vision, and Claude Code review upgrades. Prompt behavior, tokenization, and effort defaults changed enough that existing harnesses may need retuning.

GPT-Rosalind introduces life sciences reasoning in trusted-access preview
OpenAI launched GPT-Rosalind for biology, drug discovery, and translational medicine, plus a life sciences plugin for Codex. Access starts as a trusted preview for qualified customers, so near-term use is limited to partner and enterprise workflows.

OpenClaw 2026.4.15 adds Opus 4.7 support and bounded memory reads
OpenClaw 2026.4.15 adds Anthropic Opus 4.7, bundled Gemini TTS, bounded memory reads, and transport self-heal fixes. The release targets context and reliability issues users had been reporting this week.


Claude Opus 4.7 releases with xhigh effort, /ultrareview, and 3x vision resolution
Claude Opus 4.7 is now generally available across Claude, the API, and major clouds with xhigh effort, higher-resolution vision, and Claude Code review upgrades. Prompt behavior, tokenization, and effort defaults changed enough that existing harnesses may need retuning.

Qwen3.6-35B-A3B releases Apache 2.0 sparse MoE with 3B active params
Alibaba open-sourced Qwen3.6-35B-A3B, a 35B multimodal sparse MoE with only 3B active parameters under Apache 2.0. Same-day support from vLLM, Ollama, SGLang, and GGUF builders makes it immediately usable for local and production coding workloads.

Codex adds background computer use on macOS with 90+ plugins and SSH devboxes
OpenAI expanded Codex with background Mac computer use, an in-app browser, image generation, memory preview, automations, and 90+ plugins. The release moves Codex from terminal coding toward long-running UI and ops workflows, though some features remain macOS-first or alpha.

GPT-Rosalind introduces life sciences reasoning in trusted-access preview
OpenAI launched GPT-Rosalind for biology, drug discovery, and translational medicine, plus a life sciences plugin for Codex. Access starts as a trusted preview for qualified customers, so near-term use is limited to partner and enterprise workflows.
Claude Code raises Opus 4.7 subscriber limits after token burn increases
Perplexity launches Personal Computer for Mac with local file and app control
Hermes Agent launches Tool Gateway with 300+ models and bundled tools
OpenClaw 2026.4.15 adds Opus 4.7 support and bounded memory reads
Top storiesthis week
Claude Code users report 5-minute cache TTL and quota-meter regressions after March updates
GitHub issues and Hacker News threads added fresh evidence that Claude Code sessions still burn quota unexpectedly after the cache TTL change, with some users seeing usage before a prompt is sent and others recovering capacity by rolling back to 2.1.34. Watch cache reuse and metering behavior closely if you rely on long-running sessions.


Gemini 3.1 Flash TTS launches with Audio Tags, 70+ languages and API preview
Google released Gemini 3.1 Flash TTS with inline Audio Tags, multi-speaker control and 70+ languages, and opened preview access through the Gemini API and AI Studio with rollout to Vertex AI and Google Vids. Independent evals ranked it near the top of current speech leaderboards, but it runs slower and costs more than the leading system.

OpenAI Agents SDK adds sandbox execution and memory controls with Vercel, Modal, E2B and Daytona
OpenAI updated the Agents SDK with sandbox execution, memory controls and run snapshotting, and launch partners Vercel, Modal, E2B and Daytona shipped integrations. Long-running agents can now keep files, credentials and execution state in isolated runtimes instead of wiring harness, compute and storage layers together manually.

Parcae claims 1.3B Transformer quality from a 770M looped model
Together AI and UCSD released Parcae, a looped model that reuses layers with a constrained recurrent dynamic and reports stronger results than parameter-matched Transformers from 140M to 1.3B scales. The released models and code suggest recurrence can trade memory for quality under fixed FLOP budgets instead of scaling parameters alone.

Windsurf 2.0 integrates Devin for cloud agents that keep running after the IDE closes
Windsurf 2.0 launched with Devin embedded into the product, combining local agents with cloud agents that can continue across codebases after you close the laptop. The IDE now acts as a handoff layer between interactive edits and long-running remote execution.







