• bitcoinBitcoin(BTC)$77,651.00-1.40%
  • ethereumEthereum(ETH)$2,304.33-3.69%
  • tetherTether(USDT)$1.000.00%
  • rippleXRP(XRP)$1.43-0.99%
  • binancecoinBNB(BNB)$634.40-1.63%
  • usd-coinUSDC(USDC)$1.000.00%
  • solanaSolana(SOL)$85.23-2.70%
  • tronTRON(TRX)$0.328850-0.03%
  • Figure HelocFigure Heloc(FIGR_HELOC)$1.041.65%
  • dogecoinDogecoin(DOGE)$0.095884-1.02%
  • whitebitWhiteBIT Coin(WBT)$54.90-2.24%
  • USDSUSDS(USDS)$1.000.00%
  • HyperliquidHyperliquid(HYPE)$41.170.09%
  • leo-tokenLEO Token(LEO)$10.280.51%
  • bitcoin-cashBitcoin Cash(BCH)$455.97-1.82%
  • cardanoCardano(ADA)$0.245849-2.84%
  • moneroMonero(XMR)$372.15-1.67%
  • chainlinkChainlink(LINK)$9.22-2.25%
  • stellarStellar(XLM)$0.175808-1.61%
  • CantonCanton(CC)$0.150182-1.74%
  • MemeCoreMemeCore(M)$4.281.24%
  • zcashZcash(ZEC)$329.641.71%
  • daiDai(DAI)$1.00-0.02%
  • USD1USD1(USD1)$1.00-0.06%
  • litecoinLitecoin(LTC)$55.34-1.42%
  • avalanche-2Avalanche(AVAX)$9.26-2.54%
  • Ethena USDeEthena USDe(USDE)$1.00-0.01%
  • hedera-hashgraphHedera(HBAR)$0.090050-1.90%
  • suiSui(SUI)$0.93-3.38%
  • shiba-inuShiba Inu(SHIB)$0.000006-2.35%
  • RainRain(RAIN)$0.007468-3.87%
  • paypal-usdPayPal USD(PYUSD)$1.000.02%
  • the-open-networkToncoin(TON)$1.35-1.93%
  • crypto-com-chainCronos(CRO)$0.069767-0.74%
  • Circle USYCCircle USYC(USYC)$1.12-0.09%
  • tether-goldTether Gold(XAUT)$4,686.67-0.77%
  • BlackRock USD Institutional Digital Liquidity FundBlackRock USD Institutional Digital Liquidity Fund(BUIDL)$1.000.00%
  • World Liberty FinancialWorld Liberty Financial(WLFI)$0.076964-3.02%
  • Global DollarGlobal Dollar(USDG)$1.000.01%
  • BittensorBittensor(TAO)$242.25-1.80%
  • pax-goldPAX Gold(PAXG)$4,688.98-0.85%
  • mantleMantle(MNT)$0.64-1.64%
  • uniswapUniswap(UNI)$3.24-4.90%
  • polkadotPolkadot(DOT)$1.22-5.22%
  • SkySky(SKY)$0.0850771.77%
  • nearNEAR Protocol(NEAR)$1.39-2.02%
  • Falcon USDFalcon USD(USDF)$1.00-0.11%
  • okbOKB(OKB)$83.52-1.75%
  • Pi NetworkPi Network(PI)$0.167186-1.25%
  • HTX DAOHTX DAO(HTX)$0.0000020.15%
TradePoint.io
  • Main
  • AI & Technology
  • Stock Charts
  • Market & News
  • Business
  • Finance Tips
  • Trade Tube
  • Blog
  • Shop
No Result
View All Result
TradePoint.io
No Result
View All Result

Qwen Team Open-Sources Qwen3.6-35B-A3B: A Sparse MoE Vision-Language Model with 3B Active Parameters and Agentic Coding Capabilities

April 17, 2026
in AI & Technology
Reading Time: 6 mins read
A A
Qwen Team Open-Sources Qwen3.6-35B-A3B: A Sparse MoE Vision-Language Model with 3B Active Parameters and Agentic Coding Capabilities
ShareShareShareShareShare

The open-source AI landscape has a new entry worth paying attention to. The Qwen team at Alibaba has released Qwen3.6-35B-A3B, the first open-weight model from the Qwen3.6 generation, and it is making a compelling argument that parameter efficiency matters far more than raw model size. With 35 billion total parameters but only 3 billion activated during inference, this model delivers agentic coding performance competitive with dense models that are ten times its active size.

What is a Sparse MoE Model, and Why Does it Matter Here?

A Mixture of Experts (MoE) model does not run all of its parameters on every forward pass. Instead, the model routes each input token through a small subset of specialized sub-networks called ‘experts.’ The rest of the parameters sit idle. This means you can have an enormous total parameter count while keeping inference compute — and therefore inference cost and latency — proportional only to the active parameter count.

YOU MAY ALSO LIKE

Talking to AI agents is one thing — what about when they talk to each other? New startup BAND debuts ‘universal orchestrator’

Turkey wants to ban social media for kids under 15

Qwen3.6-35B-A3B is a Causal Language Model with Vision Encoder, trained through both pre-training and post-training stages, with 35 billion total parameters and 3 billion activated. Its MoE layer contains 256 experts in total, with 8 routed experts and 1 shared expert activated per token.

The architecture introduces an unusual hidden layout worth understanding: the model uses a pattern of 10 blocks, each consisting of 3 instances of (Gated DeltaNet → MoE) followed by 1 instance of (Gated Attention → MoE). Across 40 total layers, the Gated DeltaNet sublayers handle linear attention — a computationally cheaper alternative to standard self-attention — while the Gated Attention sublayers use Grouped Query Attention (GQA), with 16 attention heads for Q and only 2 for KV, significantly reducing KV-cache memory pressure during inference. The model supports a native context length of 262,144 tokens, extensible up to 1,010,000 tokens using YaRN (Yet another RoPE extensioN) scaling.

Agentic Coding is Where This Model Gets Serious

On SWE-bench Verified — the canonical benchmark for real-world GitHub issue resolution — Qwen3.6-35B-A3B scores 73.4, compared to 70.0 for Qwen3.5-35B-A3B and 52.0 for Gemma4-31B. On Terminal-Bench 2.0, which evaluates an agent completing tasks inside a real terminal environment with a three-hour timeout, Qwen3.6-35B-A3B scores 51.5 — the highest among all compared models, including Qwen3.5-27B (41.6), Gemma4-31B (42.9), and Qwen3.5-35B-A3B (40.5).

Frontend code generation shows the sharpest improvement. On QwenWebBench, an internal bilingual front-end code generation benchmark covering seven categories including Web Design, Web Apps, Games, SVG, Data Visualization, Animation, and 3D, Qwen3.6-35B-A3B achieves a score of 1397 — well ahead of Qwen3.5-27B (1068) and Qwen3.5-35B-A3B (978).

On STEM and reasoning benchmarks, the numbers are equally striking. Qwen3.6-35B-A3B scores 92.7 on AIME 2026 (the full AIME I & II), and 86.0 on GPQA Diamond — a graduate-level scientific reasoning benchmark — both competitive with much larger models.

Multimodal Vision Performance

Qwen3.6-35B-A3B is not a text-only model. It ships with a vision encoder and handles image, document, video, and spatial reasoning tasks natively.

On MMMU (Massive Multi-discipline Multimodal Understanding), a benchmark that tests university-level reasoning across images, Qwen3.6-35B-A3B scores 81.7, outperforming Claude-Sonnet-4.5 (79.6) and Gemma4-31B (80.4). On RealWorldQA, which tests visual understanding in real-world photographic contexts, the model achieves 85.3, ahead of Qwen3.5-27B (83.7) and significantly above Claude-Sonnet-4.5 (70.3) and Gemma 4-31B (72.3).

Spatial intelligence is another area of measurable gain. On ODInW13, an object detection benchmark, Qwen3.6-35B-A3B scores 50.8, up from 42.6 for Qwen3.5-35B-A3B. For video understanding, it achieves 83.7 on VideoMMMU, outperforming Claude-Sonnet-4.5 (77.6) and Gemma4-31B (81.6).

https://qwen.ai/blog?id=qwen3.6-35b-a3b

Thinking Mode, Non-Thinking Mode, and a Key Behavioral Change

One of the more practically useful design decisions in Qwen3.6 is explicit control over the model’s reasoning behavior. Qwen3.6 models operate in thinking mode by default, generating reasoning content enclosed within <think> tags before producing the final response. Developers who need faster, direct responses can disable this via an API parameter — setting "enable_thinking": False in the chat template kwargs. However, AI professionals migrating from Qwen3 should note an important behavioral change: Qwen3.6 does not officially support the soft switch of Qwen3, i.e., /think and /nothink. Mode switching must be done through the API parameter rather than inline prompt tokens.

The more novel addition is a feature called Thinking Preservation. By default, only the thinking blocks generated for the latest user message are retained; Qwen3.6 has been additionally trained to preserve and leverage thinking traces from historical messages, which can be enabled by setting the preserve_thinking option. This capability is particularly beneficial for agent scenarios, where maintaining full reasoning context can enhance decision consistency, reduce redundant reasoning, and improve KV cache utilization in both thinking and non-thinking modes.

Key Takeaways

  • Qwen3.6-35B-A3B is a sparse Mixture of Experts model with 35 billion total parameters but only 3 billion activated at inference time, making it significantly cheaper to run than its total parameter count suggests — without sacrificing performance on complex tasks.
  • The model’s agentic coding capabilities are its strongest suit, with a score of 51.5 on Terminal-Bench 2.0 (the highest among all compared models), 73.4 on SWE-bench Verified, and a dominant 1,397 on QwenWebBench covering frontend code generation across seven categories including Web Apps, Games, and Data Visualization.
  • Qwen3.6-35B-A3B is a natively multimodal model, supporting image, video, and document understanding out of the box, with scores of 81.7 on MMMU, 85.3 on RealWorldQA, and 83.7 on VideoMMMU — outperforming Claude-Sonnet-4.5 and Gemma4-31B on each of these.
  • The model introduces a new Thinking Preservation feature that allows reasoning traces from prior conversation turns to be retained and reused across multi-step agent workflows, reducing redundant reasoning and improving KV cache efficiency in both thinking and non-thinking modes.
  • Released under Apache 2.0, the model is fully open for commercial use and is compatible with the major open-source inference frameworks — SGLang, vLLM, KTransformers, and Hugging Face Transformers — with KTransformers specifically enabling CPU-GPU heterogeneous deployment for resource-constrained environments.

Check out the Technical details and Model Weights. Also, feel free to follow us on Twitter and don’t forget to join our 130k+ ML SubReddit and Subscribe to our Newsletter. Wait! are you on telegram? now you can join us on telegram as well.

Need to partner with us for promoting your GitHub Repo OR Hugging Face Page OR Product Release OR Webinar etc.? Connect with us

The post Qwen Team Open-Sources Qwen3.6-35B-A3B: A Sparse MoE Vision-Language Model with 3B Active Parameters and Agentic Coding Capabilities appeared first on MarkTechPost.

Credit: Source link

ShareTweetSendSharePin

Related Posts

Talking to AI agents is one thing — what about when they talk to each other? New startup BAND debuts ‘universal orchestrator’
AI & Technology

Talking to AI agents is one thing — what about when they talk to each other? New startup BAND debuts ‘universal orchestrator’

April 23, 2026
Turkey wants to ban social media for kids under 15
AI & Technology

Turkey wants to ban social media for kids under 15

April 23, 2026
Aevex CEO Speaks on Raising 0 Million in US IPO
AI & Technology

Aevex CEO Speaks on Raising $320 Million in US IPO

April 23, 2026
Trump Says ‘Highly Unlikely’ He Extends Iran Ceasefire
AI & Technology

Trump Says ‘Highly Unlikely’ He Extends Iran Ceasefire

April 23, 2026
Next Post
Stop using your Casely Power Pods wireless charger immediately

Stop using your Casely Power Pods wireless charger immediately

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Search

No Result
View All Result
LIVE: Artemis II crew holds a news conference on their mission to the moon | NBC News

LIVE: Artemis II crew holds a news conference on their mission to the moon | NBC News

April 23, 2026
NASA restarts work to support Europe’s uncrewed trip to Mars after years of setbacks

NASA restarts work to support Europe’s uncrewed trip to Mars after years of setbacks

April 17, 2026
Call Options Explained (Options Trading for Beginners)

Call Options Explained (Options Trading for Beginners)

April 22, 2026

About

Learn more

Our Services

Legal

Privacy Policy

Terms of Use

Bloggers

Learn more

Article Links

Contact

Advertise

Ask us anything

©2020- TradePoint.io - All rights reserved!

Tradepoint.io, being just a publishing and technology platform, is not a registered broker-dealer or investment adviser. So we do not provide investment advice. Rather, brokerage services are provided to clients of Tradepoint.io by independent SEC-registered broker-dealers and members of FINRA/SIPC. Every form of investing carries some risk and past performance is not a guarantee of future results. “Tradepoint.io“, “Instant Investing” and “My Trading Tools” are registered trademarks of Apperbuild, LLC.

This website is operated by Apperbuild, LLC. We have no link to any brokerage firm and we do not provide investment advice. Every information and resource we provide is solely for the education of our readers. © 2020 Apperbuild, LLC. All rights reserved.

No Result
View All Result
  • Main
  • AI & Technology
  • Stock Charts
  • Market & News
  • Business
  • Finance Tips
  • Trade Tube
  • Blog
  • Shop

© 2023 - TradePoint.io - All Rights Reserved!