OpenAI releases GPT‑5.2-Codex, with improvements on long-horizon work through context compaction, stronger performance on large code changes, and more
What excites me most is the leverage this gives developers. … Frederic Lardinois : At this point, I'd almost be disappointed if nobody releases a new model on Christmas Day. The current release cycle...
Sources: Mira Murati's Thinking Machines Lab is in early talks to raise funding at a ~$50B valuation, more than quadruple from July and could rise to $55B-$60B
Yeah, okay. — This bubble is running on fumes. Tim Kellogg / @timkellogg.me : i wish my blog posts paid that well [embedded post] Casey Newton / @caseynewton : If they hadn't released a product I th...
DeepSeek releases DeepSeek-OCR, a vision language model designed for efficient vision-text compression, enabling longer contexts with less compute
the new frontier of OCR from @deepseek_ai , exploring optical context compression for LLMs, is running blazingly fast on vLLM ⚡ (~2500 tokens/s on A100-40G) — powered by vllm==0.8.5 for day-0 model su...
Alibaba releases the Qwen3-VL vision models, the Qwen3Guard “safety moderation” models, and three closed-weight models, including Qwen3-Max with 1T+ parameters
Qwen 50.6k — Safetensors qwen3_vl_moe Julian Nabil / Forbes Middle East : Alibaba Introduces Qwen3-Max AI Model With Over 1T Parameters Markus Kasanmascheff / WinBuzzer : Alibaba Releases Qwen3-VL O...
Sources: Microsoft will use Anthropic's models for some AI features in its Office 365 apps, after finding Claude Sonnet 4 beats OpenAI's GPT-5 in some tasks
Microsoft moves beyond OpenAI? Igor Bonifacic / Engadget : Microsoft reportedly plans to start using Anthropic models to power some of Office 365's Copilot features Sherwood News : Census data shows d...
DeepSeek details V3.1 and says it surpasses R1 on key benchmarks and is customized to work with next-gen Chinese-made AI chips, after unveiling it on August 19
Introducing DeepSeek-V3.1: our first step toward the agent era! 🚀 Tobias Mann / The Register : DeepSeek's new V3.1 release points to potent new Chinese chips coming soon Hugging Face : DeepSeek-V3.1 ...
OpenAI releases gpt-oss-120b and gpt-oss-20b, its first open-weight models since GPT-2; the smaller gpt-oss-20b can run locally on a device with 16GB+ of RAM
gpt-oss-120b and gpt-oss-20b push the frontier of open-weight reasoning models Simon Willison / Simon Willison's Weblog : OpenAI's new open weight (Apache 2) models are really good OpenAI on GitHub : ...
Cerebras announces the $50/month Code Pro and the $200/month Code Max plans, offering users access to Qwen3-Coder at speeds of up to 2,000 tokens per second
Two interesting examples of inference speed as a flagship feature of LLM services today. Bluesky: Tim Kellogg / @timkellogg.me : Cerebras Code — use models hosted on Cerebras with a dev-friendly subsc...
Z.ai, formerly known as Zhipu and that has raised $1.5B from Tencent and others, releases GLM-4.5, an open-source AI model that it says is cheaper than DeepSeek
chinese models really are taking over huh Simon Willison / @simonwillison.net : Pretty decent pelicans from the new GLM-4.5 and GLM-4.5 Air models. Both models are MIT licensed, released by Chinese A...
Alibaba debuts the Qwen3-Coder model for agentic coding, including a 480B-parameter MoE variant, and open sources Qwen Code, a CLI tool adapted from Gemini CLI
Qwen 39.4k — Text Generation Transformers Safetensors qwen3_moe conversational Coco Feng / South China Morning Post : Alibaba upgrades flagship Qwen3 model to outperform OpenAI, DeepSeek in maths, c...