Qwen 3.5-35B-A3B Surpasses GPT-OSS-120B as Daily Driver at 1/3 the Size

Original: Qwen 3.5-35B-A3B is beyond expectations. It's replaced GPT-OSS-120B as my daily driver and it's 1/3 the size. View original →

Read in other languages: 한국어日本語
LLM Mar 1, 2026 By Insights AI (Reddit) 1 min read 3 views Source

Qwen 3.5-35B-A3B Exceeds All Expectations

The LocalLLaMA community is rallying around Alibaba's Qwen 3.5-35B-A3B, with a highly-upvoted post declaring it has replaced GPT-OSS-120B as the poster's daily driver — at just one-third the size.

MoE Efficiency: 35B Parameters, 3B Active

The model employs a Mixture of Experts (MoE) architecture: 35 billion total parameters, but only approximately 3 billion are active during inference. This makes it dramatically cheaper to run than comparable dense models while maintaining high output quality.

Real-World Performance

The original poster shared specific production use cases where the model excels:

  • Automated message and email triage via N8N with priority-based batching
  • Agent systems with dynamic tool selection
  • General-purpose development assistance

The consensus in the thread is that Qwen 3.5-35B-A3B punches well above its weight class, particularly for coding and reasoning tasks.

Qwen 3.5 Family Context

The Qwen 3.5 series is Alibaba's latest open-source model family, available in sizes ranging from compact dense models to larger MoE variants. The 35B-A3B continues a trend of Chinese open-source models closing the gap with — and in some cases surpassing — Western counterparts at comparable parameter counts.

Share:

Related Articles

LLM Reddit Mar 2, 2026 1 min read

Alibaba's Qwen team has released Qwen 3.5 Small, a new small dense model in their flagship open-source series. The announcement topped r/LocalLLaMA with over 1,000 upvotes, reflecting the local AI community's enthusiasm for capable small models.

LLM Reddit Mar 2, 2026 1 min read

A remarkable 13-month comparison: running frontier-level DeepSeek R1 at ~5 tokens/second cost $6,000 in early 2025. Today, you can run a significantly stronger model at the same speed on a $600 mini PC — and get 17-20 t/s with even more capable models.

Comments (0)

No comments yet. Be the first to comment!

Leave a Comment

© 2026 Insights. All rights reserved.