2026-02-12
This is a HUGE win for developers. Claude Code is excellent, but the $200/mo Max plan can be expensive for daily use. GLM-5 works inside Claude Code, with (arguably) comparable performance at ~1/3 the cost. Setup takes ~1 minute: • Install Claude Code as usual • Run 'npx
Z.ai
Z.ai launches GLM-5, saying its flagship open-weight model has “best-in-class performance among all open-source models” in reasoning, coding, and agentic tasks
We are launching GLM-5, targeting complex systems engineering and long-horizon agentic tasks. Scaling is still one of the most important ways …
This is a HUGE win for developers. Claude Code is excellent, but the $200/mo Max plan can be expensive for daily use. GLM-5 works inside Claude Code, with (arguably) comparable performance at ~1/3 the cost. Setup takes ~1 minute: • Install Claude Code as usual • Run 'npx
Reuters
Z.ai says it will raise prices by at least 30% for new GLM coding plan subscribers to accommodate surging demand for its AI coding tools
2026-02-11
This is a HUGE win for developers. Claude Code is excellent, but the $200/mo Max plan can be expensive for daily use. GLM-5 works inside Claude Code, with (arguably) comparable performance at ~1/3 the cost. Setup takes ~1 minute: • Install Claude Code as usual • Run 'npx
Z.ai
Z.ai launches GLM-5, its flagship open-weight model, saying it has best-in-class performance among open-source models in reasoning, coding, and agentic tasks
We are launching GLM-5, targeting complex systems engineering and long-horizon agentic tasks. Scaling is still one of the most important ways …
2025-12-21
One thing I really like about this CoT monitorability is the reframing that CoT isn't a truth oracle, but more like a control interface. The point isn't that CoT faithfully reflects how the model really reasons, but that it gives us signals we can observe, poke at, and use to
OpenAI
OpenAI introduces a framework to evaluate chain-of-thought monitorability and a suite of 13 evaluations designed to measure the monitorability of an AI system
2025-12-20
One thing I really like about this CoT monitorability is the reframing that CoT isn't a truth oracle, but more like a control interface. The point isn't that CoT faithfully reflects how the model really reasons, but that it gives us signals we can observe, poke at, and use to
OpenAI
OpenAI introduces a framework to evaluate chain-of-thought monitorability and a suite of 13 evaluations designed to measure the monitorability of an AI system
We introduce evaluations for chain-of-thought monitorability and study how it scales with test-time compute, reinforcement learning, and pretraining.