/
Navigation
Chronicles
Browse all articles
Explore
Semantic exploration
Research
Entity momentum
Nexus
Correlations & relationships
Story Arc
Topic evolution
Drift Map
Semantic trajectory animation
Posts
Analysis & commentary
Pulse API
Tech news intelligence API
Browse
Entities
Companies, people, products, technologies
Domains
Browse by publication source
Handles
Browse by social media handle
Detection
Concept Search
Semantic similarity search
High Impact Stories
Top coverage by position
Sentiment Analysis
Positive/negative coverage
Anomaly Detection
Unusual coverage patterns
Analysis
Rivalry Report
Compare two entities head-to-head
Semantic Pivots
Narrative discontinuities
Crisis Response
Event recovery patterns
Connected
Search: /
Command: ⌘K
Embeddings: large
TEXXR

Chronicles

The story behind the story

days · browse · Enter similar · o open

Alibaba launches Qwen3.6-27B, an open-weight dense model with 27B parameters, saying it surpasses Qwen3.5-397B-A17B on major coding benchmarks

· 4226 words  · QwenTeam丨Translations:.体中文  —  HUGGING FACE  —  MODELSCOPE  —  DISCORD

Qwen

Discussion

  • @suchenzang Susan Zhang on x
    “china is becoming more closed-source” - cope
  • @onlyterp Terp on x
    ok.... So this just happened Qwen 3.6 27b running locally on my 5090 straight up beating mimo v2.5 pro 😭 [image]
  • @ollama @ollama on x
    Qwen 3.6 27B model is available on Ollama! Use it with all the integrations in Ollama or chat with the model. Chat with the model: ollama run qwen3.6:27b OpenClaw: ollama launch openclaw —model qwen3.6:27b Claude Code: ollama launch claude —model qwen3.6:27b More
  • @sudoingx @sudoingx on x
    okay this is absolutely insane. my undisputed king qwen 3.5-27b dense on single RTX 3090 just got replaced by the same team today. qwen drops 3.6-27b dense just now and the chart says it beats its predecessor on every single benchmark, beats qwen 3.5-397b-a17b moe which is 15x [i…
  • @alibaba_qwen @alibaba_qwen on x
    🚀 Meet Qwen3.6-27B, our latest dense, open-source model, packing flagship-level coding power! Yes, 27B, and Qwen3.6-27B punches way above its weight. 👇 What's new: 🧠 Outstanding agentic coding — surpasses Qwen3.5-397B-A17B across all major coding benchmarks 💡 Strong [image]
  • @cgtwts @cgtwts on x
    Qwen just dropped Qwen3.6-27B >open source >a dense 27b model >beats their own 397B flagship on coding >14x smaller and easier to run >strong at agentic coding >handles both text and images >has fast mode and deep thinking mode >much cheaper to run locally [video]
  • @kylehessling1 Kyle Hessling on x
    Guys, I am absolutely astounded. The Qwen 3.6 27b is like a jump to Qwen 4 from Qwen 27B 3.5. I just did a full suite of front end design tests and agentic benchmarks, made entirely by it. VERDICT: They're so much better than I thought they'd be, like I'm completely astounded. I
  • @alibaba_qwen @alibaba_qwen on x
    LM Performance:With only 27B parameters, Qwen3.6-27B outperforms the Qwen3.5-397B-A17B (397B total / 17B active, ~15x larger!) on every major coding benchmark — including SWE-bench Verified (77.2 vs. 76.2), SWE-bench Pro (53.5 vs. 50.9), Terminal-Bench 2.0 (59.3 vs. 52.5), and [i…
  • @alibabagroup @alibabagroup on x
    🚀 Qwen3.6-27B is now open source! Start building with this dense 27B multimodal model delivering flagship-level agentic coding performance. #AlibabaAI #Qwen
  • @hxiao Han Xiao on x
    With 3.6-27b release, the dense-over-MoE gap is shrinking, which is good for local AI as MoE like 35b-a3b are more friendly on low-budget GPU and support much longer context (256k full easily on 24gb vram). Same-scale comparison (27B dense vs 35B-A3B MoE): dense still wins most […
  • @alibaba_qwen @alibaba_qwen on x
    VLM Performance:Qwen3.6-27B is natively multimodal, supporting both vision-language thinking and non-thinking modes in a single unified checkpoint — the same as Qwen3.6-35B-A3B. It handles images and video alongside text, enabling multimodal reasoning, document understanding, [im…
  • @sudoingx @sudoingx on x
    this was supposed to be a normal evening, then i saw on the timeline that qwen 3.6 27b dense q4 weights from unsloth are live and i could not sit still. compiled llama.cpp with cuda on the single rtx 3090 at 2am from bangkok, launched with the exact same flags that crowned [image…
  • r/LocalLLaMA r on reddit
    Qwen3.6-27B released!