Skip to content
AI Primer

Explore what's new in AI

Where people deep in AI come to stay current.

Filters

Category

Tags

GPT-5.5 users report 4-10x shorter runs and smoother tool calls one day after launch
New

GPT-5.5 users report 4-10x shorter runs and smoother tool calls one day after launch

Users and third-party evals reported shorter runs, stronger long-context scores, and faster rollout into Cursor and other tools a day after GPT-5.5 hit the API. Higher per-token pricing may be partly offset by lower loop time and fewer tool-call stalls, so watch early bench data before changing defaults.

Codex25th April·5 min read
DeepSeek cuts V4-Pro API 75% to $0.43/$0.87 per 1M tokens through May 5
New

DeepSeek cuts V4-Pro API 75% to $0.43/$0.87 per 1M tokens through May 5

DeepSeek lowered V4-Pro API pricing and updated integration guidance for Claude Code, OpenCode, and OpenClaw a day after V4 launched. Check whether V4-Flash is the easier deploy today, while Pro stays heavier and more rate-limited.

LLM Serving25th April·6 min read
Breaking

Qwen-Image-2.0-Pro launches at #9 on Arena with multilingual text rendering

Alibaba launched Qwen-Image-2.0-Pro on ModelScope and API with better prompt adherence, multilingual typography, and steadier style quality. The model is aimed at text-heavy jobs like UI mockups and posters, so test it for layout-heavy generation.

Qwen-Image-2.0-Pro launches at #9 on Arena with multilingual text rendering
New
Qwen·25th April·3 min read
See all stories →
🤖Agentic Engineering(25)
🧩Agent Development(3)
🧠Models & APIs(2)
Inference & Infrastructure(3)
🔒Security & Reliability(1)
📊Business & Policy(2)

Top storiesthis week

Release

DeepSeek V4 reports CSA/HCA attention and 10% KV cache at 1M context

Engineers unpacked DeepSeek V4's hybrid CSA/HCA attention a day after launch; it claims 27% of V3.2 FLOPs and 10% of its KV cache at 1M tokens. External tests pushed V4 Pro near the top of open-model indexes, but users also reported rate limits and mixed third-party results.

DeepSeek V4 reports CSA/HCA attention and 10% KV cache at 1M context
Inference Optimization·24th April·8 min read
See all stories →
AI PrimerAI Primer

Your daily guide to AI tools, workflows, and creative inspiration.

© 2026 AI Primer. All rights reserved.