Alibaba releases its open-weight Qwen3.5 Small Model Series in 0.8B, 2B, 4B, and 9B sizes, claiming the 9B model rivals OpenAI's gpt-oss-120b on some benchmarks
Earlier today, e-commerce giant Alibaba's Qwen Team of AI researchers, focused primarily on developing and releasing to the world …
Alibaba releases the open-weight Qwen3.5 Small Model Series in 0.8B, 2B, 4B, and 9B sizes, claiming the 9B model rivals OpenAI's gpt-oss-120b on some benchmarks
Z.ai launches GLM-5, saying its flagship open-weight model has “best-in-class performance among all open-source models” in reasoning, coding, and agentic tasks
We are launching GLM-5, targeting complex systems engineering and long-horizon agentic tasks. Scaling is still one of the most important ways …
Z.ai says it will raise prices by at least 30% for new GLM coding plan subscribers to accommodate surging demand for its AI coding tools
Mistral launches Mistral 3, a family of 10 models under the Apache 2.0 license, including its new flagship Mistral Large 3 and nine smaller Ministral 3 models
Mistral AI, Europe's most prominent artificial intelligence startup, is releasing its most ambitious product suite to date …
Nvidia says it will begin selling the DGX Spark mini PC, with DGX OS, for AI developers on October 15 on Nvidia.com and select third-party retailers for $3,999
(PCMag/Michael Kan) … It's not a consumer desktop, but Nvidia's foray into an AI developer-focused mini PC is finally ready to launch.
Google announces Gemma 3 270M, a compact model designed for task-specific fine-tuning with strong capabilities in instruction following and text structuring
ai.google.dev/gemma/docs/c... Tim Duffy / @timfduffy.com : Google just released a 270M parameter Gemma model. As a tiny model lover I'm excited. Models in this size class are usu...
Alibaba debuts its Qwen3 family of open-weight “hybrid” AI reasoning models, including Qwen3-235B-A22B, with 235B total parameters and 22B activated parameters
Chinese tech company Alibaba on Monday released Qwen3, a family of AI models the company claims matches …
Mistral launches Small 3, a latency-optimized 24B-parameter model that it says is competitive with larger models such as Llama 3.3 70B or Qwen 32B
Apache 2.0, 81% MMLU, 150 tokens/s — Today we're introducing Mistral Small 3, a latency-optimized 24B-parameter model released under the Apache 2.0 license.
Alibaba releases 32.5B-parameter QwQ-32B-Preview under Apache 2.0 and claims the “reasoning” AI model beats OpenAI's o1-preview on the AIME and MATH tests
Introduction QwQ-32B-Preview is an experimental research model developed … Ananya Gairola / Benzinga : Alibaba's New AI Model Outperforms OpenAI's o1 In Specific Benchmarks, Now Av...
Mistral AI releases 22B-parameter Codestral, its first generative AI model for coding, trained on 80+ programming languages and prohibited for commercial use
You need to agree to share your contact information to access this model Deepti Pathak / Fossbytes : Mistral AI Launches Codestral: AI Code Generation Across 80 Programming Languag...
Meta releases Code Llama 70B, a new version of its code generation model, featuring improved code correctness, a variant optimized for Python, and more
available under the same license as previous Code Llama models. Download the models ➡️ https://ai.meta.com/... • CodeLlama-70B • CodeLlama-70B-Python • CodeLlama-70B-Instruct [imag...
Meta releases Code Llama 70B, a new version of its code generation model, featuring improved code correctness, a variant optimized for Python, and more
available under the same license as previous Code Llama models. Download the models ➡️ https://ai.meta.com/... • CodeLlama-70B • CodeLlama-70B-Python • CodeLlama-70B-Instruct [imag...