Fresh stories

GPT-5.5 users report 4-10x shorter runs and smoother tool calls one day after launch
Users and third-party evals reported shorter runs, stronger long-context scores, and faster rollout into Cursor and other tools a day after GPT-5.5 hit the API. Higher per-token pricing may be partly offset by lower loop time and fewer tool-call stalls, so watch early bench data before changing defaults.

DeepSeek cuts V4-Pro API 75% to $0.43/$0.87 per 1M tokens through May 5
DeepSeek lowered V4-Pro API pricing and updated integration guidance for Claude Code, OpenCode, and OpenClaw a day after V4 launched. Check whether V4-Flash is the easier deploy today, while Pro stays heavier and more rate-limited.
Qwen-Image-2.0-Pro launches at #9 on Arena with multilingual text rendering
Alibaba launched Qwen-Image-2.0-Pro on ModelScope and API with better prompt adherence, multilingual typography, and steadier style quality. The model is aimed at text-heavy jobs like UI mockups and posters, so test it for layout-heavy generation.


GPT-5.5 users report 4-10x shorter runs and smoother tool calls one day after launch
Users and third-party evals reported shorter runs, stronger long-context scores, and faster rollout into Cursor and other tools a day after GPT-5.5 hit the API. Higher per-token pricing may be partly offset by lower loop time and fewer tool-call stalls, so watch early bench data before changing defaults.

DeepSeek cuts V4-Pro API 75% to $0.43/$0.87 per 1M tokens through May 5
DeepSeek lowered V4-Pro API pricing and updated integration guidance for Claude Code, OpenCode, and OpenClaw a day after V4 launched. Check whether V4-Flash is the easier deploy today, while Pro stays heavier and more rate-limited.

Claude Code users report 30-40% token growth and incomplete long tasks
Users reported higher token use, partial long-document reviews, and rising spend on routine tasks after Claude Code regressions came into focus. Some developers still get strong results in constrained harnesses, but others may want to switch to Codex for long-running work.

ClawSweeper closes 4,000 OpenClaw issues with 50 Codex agents in one day
Steipete’s maintainer bot ran 50 Codex agents in parallel and closed about 4,000 OpenClaw issues in a day. The cleanup pushed into rate limits, so use the README dashboard and Project Clowfish clustering to track large agent sweeps.
Kilo Code opens Roo migration with --install-extension and AGENTS.md conversion
Qwen-Image-2.0-Pro launches at #9 on Arena with multilingual text rendering
OpenClaw 2026.4.24 adds voice-call handoff and browser recovery
Tool vendors add GPT-5.5 to Cursor, Databricks, Droid, and ml-intern within 24 hours
Top storiesthis week
DeepSeek V4 reports CSA/HCA attention and 10% KV cache at 1M context
Engineers unpacked DeepSeek V4's hybrid CSA/HCA attention a day after launch; it claims 27% of V3.2 FLOPs and 10% of its KV cache at 1M tokens. External tests pushed V4 Pro near the top of open-model indexes, but users also reported rate limits and mixed third-party results.


OpenAI opens GPT-5.5 API with 1M context and Responses support
OpenAI added GPT-5.5 and GPT-5.5 Pro to the API and Playground with 1M context and Responses support. Partners including OpenRouter, Perplexity, GitHub Copilot, Vercel, Warp, and Devin rolled it out the same day, widening access beyond Codex.

Sakana Fugu opens beta with OpenAI-compatible API
Sakana AI opened beta access to Fugu, a multi-agent orchestration system that routes work across multiple frontier models via an OpenAI-compatible API. The launch packages model selection and role assignment as an external runtime, but access is still gated by beta signup.

Cursor 3.2 adds /multitask async subagents, worktrees, and GPT-5.5
Cursor 3.2 added /multitask async subagents, improved worktrees, and multi-root workspaces, then paired the release with GPT-5.5 rollout at 72.8% on CursorBench. The update makes background agent orchestration a first-class IDE workflow instead of a blocking queue.

BidirLM-Omni-2.5B-Embedding launches 2048-dim text-image-audio vectors
BidirLM released a 2.5B multilingual encoder that embeds text, images, and audio into one shared 2048-dimensional space and works directly with Sentence Transformers. It tops several open-data embedding leaderboards and can run locally on GPU.








