Qwen 3.5-35B-A3B Surpasses GPT-OSS-120B as Daily Driver at 1/3 the Size
Original: Qwen 3.5-35B-A3B is beyond expectations. It's replaced GPT-OSS-120B as my daily driver and it's 1/3 the size. View original →
Qwen 3.5-35B-A3B Exceeds All Expectations
The LocalLLaMA community is rallying around Alibaba's Qwen 3.5-35B-A3B, with a highly-upvoted post declaring it has replaced GPT-OSS-120B as the poster's daily driver — at just one-third the size.
MoE Efficiency: 35B Parameters, 3B Active
The model employs a Mixture of Experts (MoE) architecture: 35 billion total parameters, but only approximately 3 billion are active during inference. This makes it dramatically cheaper to run than comparable dense models while maintaining high output quality.
Real-World Performance
The original poster shared specific production use cases where the model excels:
- Automated message and email triage via N8N with priority-based batching
- Agent systems with dynamic tool selection
- General-purpose development assistance
The consensus in the thread is that Qwen 3.5-35B-A3B punches well above its weight class, particularly for coding and reasoning tasks.
Qwen 3.5 Family Context
The Qwen 3.5 series is Alibaba's latest open-source model family, available in sizes ranging from compact dense models to larger MoE variants. The 35B-A3B continues a trend of Chinese open-source models closing the gap with — and in some cases surpassing — Western counterparts at comparable parameter counts.
Related Articles
Alibaba's Qwen team has released Qwen 3.5 Small, a new small dense model in their flagship open-source series. The announcement topped r/LocalLLaMA with over 1,000 upvotes, reflecting the local AI community's enthusiasm for capable small models.
Users on r/LocalLLaMA have spotted Qwen3.5 model names appearing in Alibaba's official Qwen chat interface, signaling an imminent release of the next generation of Alibaba's open-source LLM series.
A remarkable 13-month comparison: running frontier-level DeepSeek R1 at ~5 tokens/second cost $6,000 in early 2025. Today, you can run a significantly stronger model at the same speed on a $600 mini PC — and get 17-20 t/s with even more capable models.
Comments (0)
No comments yet. Be the first to comment!