Fresh stories
Nemotron 3 Nano Omni launches 30B-A3B multimodal model with 256K context
NVIDIA opened Nemotron 3 Nano Omni, a 30B-A3B model for text, image, audio, and video, with day-one serving support. That lets teams run one open model for perception-heavy agents instead of stitching separate components.

Poolside releases Laguna M.1 and XS.2 coding models with 225B/23B and 33B/3B MoEs
Poolside opened Laguna M.1 and Laguna XS.2 as its first public coding models, with Apache 2.0 weights and same-day provider support. That gives teams open coding models that can run locally or through standard serving stacks.

Gemini adds Grounding with Exa for websites, docs, people, and company search
Gemini models can now use Grounding with Exa to search websites, technical docs, papers, people, and companies through Exa's index. That gives Gemini a new agent-style grounding path alongside Google's first-party search tooling.


Nemotron 3 Nano Omni launches 30B-A3B multimodal model with 256K context
NVIDIA opened Nemotron 3 Nano Omni, a 30B-A3B model for text, image, audio, and video, with day-one serving support. That lets teams run one open model for perception-heavy agents instead of stitching separate components.

Opus 4.7 users report verbose output, weaker 1M context, and 12–27% higher costs
Users reported more verbosity, weaker 1M-context behavior, and little coding gain after Opus 4.7 rolled out. OpenRouter measured 12–27% higher costs, and some teams reverted their default model.

Codex adds macOS computer use, in-app browser, and artifact previews
Codex gained background macOS control, page inspection, image generation, plugins, artifacts, and follow-up automations. That gives it one agent thread for desktop apps, frontend debugging, and recurring work.

Poolside releases Laguna M.1 and XS.2 coding models with 225B/23B and 33B/3B MoEs
Poolside opened Laguna M.1 and Laguna XS.2 as its first public coding models, with Apache 2.0 weights and same-day provider support. That gives teams open coding models that can run locally or through standard serving stacks.
Mistral launches Workflows public preview with durable execution and human approvals
AWS and OpenAI launch Bedrock Managed Agents with Codex and model access in limited preview
Plurai introduces vibe-training with sub-100ms agent guardrails and 43% fewer failures
Gemini adds Grounding with Exa for websites, docs, people, and company search

Sigma launches private AI browser with local OpenClaw, Gemma 4, and Qwen support

ElevenLabs releases Agent Templates with 50+ support, SDR, and training workflows

Claude Connectors add Blender and Autodesk Fusion control via MCP

Warp opens source with Oz-managed agents and public roadmap

Base44 launches platform migrations with 2-click imports from Salesforce, Shopify, and WordPress

Helmor launches local-first Conductor alternative with one-click import
Top storiesthis week
OpenClaw 2026.4.26 adds Google Live Talk, openclaw migrate, and Matrix E2EE
OpenClaw 2026.4.26 shipped Google Live Talk, local-model fixes, openclaw migrate imports for Claude and Hermes, and one-command Matrix E2EE. It also hardens plugins, Docker, and transcript compaction for self-hosted agent runs.


Bedrock adds OpenAI models and stateful runtime in coming weeks
AWS says OpenAI models will land on Bedrock in coming weeks alongside a new stateful runtime. OpenAI also said its Microsoft partnership is now non-exclusive, which opens a multi-cloud path for deployment and procurement.

Symphony launches Codex orchestration for Linear and GitHub issue queues
OpenAI released Symphony, an orchestration layer that turns issue trackers into Codex agent queues for PR generation and review. Early users say it can move many tickets in parallel, but token burn rises quickly when agents fan out.

mattpocock/skills ranks #1 on GitHub at 28K stars with `/grill-me` and `/tdd` packs
mattpocock/skills hit the top of GitHub Trending as reusable `SKILL.md` packs for grilling specs, writing PRDs, and enforcing TDD spread across coding-agent workflows. The format is starting to look like a distribution layer for agent behavior, with faster install tooling and third-party skills shipping around the same pattern.

vLLM 0.20.0 releases TurboQuant 2-bit KV cache, CUDA 13 baseline, and DeepSeek V4 upgrades
vLLM 0.20.0 shipped a new CUDA 13 / PyTorch 2.11 / Transformers v5 baseline, TurboQuant 2-bit KV cache, FA4 MLA defaults, and deeper DeepSeek V4 support. The release changes serving baselines across NVIDIA, AMD, Intel, and ARM-CUDA setups, including 4x KV capacity and a clearer upgrade path for teams already running V4.






