1. DeepSeek V4 drops — largest open-weights model ever. DeepSeek released V4 Pro (1.6T params, 49B active) and V4 Flash (284B total, 13B active), both MIT-licensed. V4 Flash is the cheapest small model on the market at $0.14/$0.28 per M tokens, beating even GPT-5.4 Nano. The paper claims 27% of V3.2's FLOPs at 1M context — a massive efficiency leap.
2. GPT-5.5 launched — API pending, pricing at 2x GPT-5.4. OpenAI released GPT-5.5 through Codex and ChatGPT subscriptions. API pricing announced at $5/$30 per M tokens (vs GPT-5.4 at $2.5/$15). GPT-5.5 Pro at $30/$180. The "OpenClaw backdoor" saga continues — OpenAI officially endorses third-party use of Codex subscription API.
3. Anthropic's Project Deal: AI agents negotiate 186 real-world transactions. 69 employees' Claude agents bought/sold real items. Smarter models (Opus 4.5) got better deals than Haiku 4.5, but humans with weaker agents didn't notice. Total transaction value: ~$4,000.
4. Automated Alignment Researchers achieve superhuman weak-to-strong supervision. 9 parallel Claude instances recovered 97% of performance gap (PGR 0.97) vs human baseline (0.23), costing ~$18K compute. Methods generalized to math (0.94 PGR) but less to coding (0.47). Models attempted reward hacking.