Fresh stories
OpenAI launches Daybreak with GPT-5.5-Cyber, Codex workflows, and repo scanning
OpenAI launched Daybreak, combining GPT-5.5, Codex workflows, repo scanning, threat modeling, and patch generation for cyber-defense teams. It packages frontier models into a continuous secure-software workflow, so teams can test whether it fits their response pipeline.

OpenAI launches Deployment Company with $4B backing and 150 forward-deployed engineers
OpenAI launched the OpenAI Deployment Company and tied it to Tomoro’s acquisition, giving the unit 150 forward-deployed engineers and $4 billion in initial backing from 19 partners. It matters because OpenAI is packaging services, deployment help, and organizational integration as part of the product stack instead of leaving enterprise rollout to outside consultancies.

Anthropic launches Claude Platform on AWS with native billing, IAM, and Managed Agents
Anthropic made Claude Platform on AWS generally available, exposing the native Claude API with AWS authentication, billing, CloudTrail, and commitment retirement. It lets teams use Managed Agents and related Claude features inside existing AWS governance workflows.


TanStack reports npm supply-chain attack across 42 packages with credential-stealing payload
TanStack disclosed a supply-chain attack that pushed two malicious npm versions across 42 packages in a 10-minute window. The payload targeted cloud keys, GitHub tokens, npm credentials, and SSH material, so teams should audit installs and rotate secrets.

OpenAI launches Daybreak with GPT-5.5-Cyber, Codex workflows, and repo scanning
OpenAI launched Daybreak, combining GPT-5.5, Codex workflows, repo scanning, threat modeling, and patch generation for cyber-defense teams. It packages frontier models into a continuous secure-software workflow, so teams can test whether it fits their response pipeline.

Thinking Machines introduces interaction models with 200 ms full-duplex audio, video, and tool use
Thinking Machines previewed interaction models that process audio, video, and text in 200 ms micro-turns, letting the system listen, speak, and react at the same time. The demos matter because the interaction loop is trained into the model instead of stitched together from separate speech and tool layers.

Claude Code 2.1.139 adds `claude agents` view and `/goal` runs with live turn counters
Claude Code 2.1.139 shipped a research-preview agent view plus a `/goal` mode that keeps working across turns while showing elapsed time, turns, and token counts. The update turns parallel Claude sessions into a built-in control plane, so teams can drop tmux-and-scripts workarounds.
OpenAI launches Deployment Company with $4B backing and 150 forward-deployed engineers
Artificial Analysis launches Coding Agent Index: Cursor plus Opus 4.7 scores 61, Codex plus GPT-5.5 60
Developers launch Markdown Experience Guidelines and HTML explainer tools for agent output
Anthropic launches Claude Platform on AWS with native billing, IAM, and Managed Agents

Developers launch Agent FM, Mate, and ntm for multi-session Claude Code and Codex control

Nous Research adds CUA computer use to Hermes Agent for desktop control

OpenBMB releases MiniCPM-V 4.6 1.3B with 75.7 ms TTFT and 19x token efficiency

Files SDK launches unified storage API with 18 providers and OpenAI, Vercel AI, and Claude tools
Top storiesthis week
Codex app adds /goal for long-running React Doctor and iOS runs
OpenAI staff said /goal is now available in the Codex app, and users posted long-running runs that fixed React Doctor scores, built iOS features, and queued weekend tasks. The update moves Codex from CLI-only planning to persistent, steerable work sessions.


Local users report DeepSeek V4 Flash, Qwen 3.6, and Gemma 4 at 40-200 tok/s on Macs and 3090s
Developers posted new local-model measurements for DS4, Qwen 3.6, and Gemma 4: about 40 tok/s on an M3 Ultra, 70+ tok/s on MacBooks with MPS, and 120-200 tok/s for Qwen3.6-27B on a single RTX 3090. The numbers suggest coding-capable local runs are moving from demos toward regular use.

DFlash adds Qwen3-8B speculator with 82.2% first-token acceptance
Posts said Qwen3-8B now has a DFlash speculator with 82.2% first-token acceptance and 3.74 accepted tokens per step, alongside broader DFlash claims of over 6x lossless acceleration. It matters because the release turns a decoding paper into a concrete speculative-inference artifact engineers can test against existing Qwen stacks.

OpenCode adds Ring 2.6 1T with 256K context and free limited-time access
OpenCode made Ring 2.6 1T available in the editor with reasoning enabled and free access for a limited period. Follow-on posts from Kilo and others claim frontier-level results on AIME 26, ClawEval, Gaia2-search, and Tau2-Bench Telecom.

GPT-5.5 vs Opus 4.7: users compare plan mode, frontend output, and 120K-context use
User posts and HN threads compared GPT-5.5 and Opus 4.7 across plan mode, frontend work, and 120K-context sessions. The split results mean token burn and instruction discipline matter as much as raw benchmark scores.






