3 Billion Active Parameters Just Challenged 30 Billion: Inside Qwen3.6’s Sparse MoE
Alibaba’s Qwen3.6-35B-A3B activates only 3B parameters per token yet claims agentic coding parity with models 10x its size. We dissect the architecture, benchmarks, and whether this Apache 2.0 release actually changes the local AI equation.