2026-04-04

Qwen 3.5 Medium Review: 7x Less RAM, Same Quality (April 2026)

Alibaba's Qwen 3.5 Medium series remains a masterclass in one thing: smart architecture beats raw parameters. And in April 2026, the Qwen ecosystem has grown even stronger. Qwen 3.5 Medium Series benchmarks Qwen 3.5: Same quality, 7x less RAM than competitors
April 2026 Update: Since this article was first published, Alibaba shipped Qwen 3.6-Plus (API-only flagship), Qwen3-Coder-Next (80B MoE coding model you can run locally), four new Qwen 3.5 Small models (0.8B-9B), and Ollama gained MLX support delivering up to 93% faster decode on Apple Silicon. Scroll to the April 2026 Update section for full details.

The 4 New Models

ModelTotal ParamsActive ParamsRAM Q4Context
Qwen3.5-Flash35B22 GB1M tokens
Qwen3.5-35B-A3B ⭐35B3B20 GB128K
Qwen3.5-27B27B27B16 GB128K
Qwen3.5-122B-A10B122B10B72 GB128K

The Revolution: Qwen3.5-35B-A3B

This model is special.

MoE Architecture: 35B total parameters, only 3B active per token
  • 35B total, but only 3B active thanks to MoE (Mixture of Experts) architecture
  • It beats Qwen3-235B-A22B — a model 7× larger
  • It runs on a MacBook Pro M3 24GB (or M4 24GB)
Translation: You get near-frontier quality on a standard laptop, without needing a Mac Studio 512GB.

Why Is This Possible?

Alibaba focused on:

  • Better architecture (smarter expert routing)
  • Data quality (less but better)
  • Optimized RL (Reinforcement Learning)

The result: More intelligence with less compute.

Performance Comparison

Qwen3.5-35B-A3B vs Qwen3-235B-A22B

Metric35B-A3B235B-A22BWinner
Quality (MMLU)82.1%79.5%35B-A3B
RAM Q4_K_M20 GB~140 GB35B-A3B
Speed (M4 Mac)~45 tok/s~15 tok/s35B-A3B
API Price$0.002/1K$0.008/1K35B-A3B
💡 The model 7× smaller is better, faster, and 4× cheaper.

Which Mac?

Qwen ModelMinimum ConfigRecommendedSpeed (pre-MLX)Speed (Ollama 0.19 MLX)
9B (Small)MacBook Air M2 8GBAny Mac 16GB~80 tok/s~120 tok/s
27BM2 16GBM3 24GB~55 tok/s~85 tok/s
35B-A3BM3 Pro 24GBMacBook Pro M4 32GB~45 tok/s~75 tok/s
FlashM3 Pro 24GBMac Studio M4 32GB~35 tok/s~55 tok/s
Coder-NextMac Studio M4 64GBMac Studio M4 Pro 96GB~30 tok/s~50 tok/s
122B-A10BMac Studio M2 Ultra 96GBMac Studio M4 Ultra 128GB~30 tok/s~45 tok/s
Updated April 2026 with Ollama 0.19 MLX speeds. Actual speeds vary by quantization and system load.

April 2026 Update: What's Changed

The Qwen ecosystem has expanded significantly since February. Here is everything that matters for Mac users running models locally.

Qwen 3.6-Plus — New API Flagship (April 2, 2026)

Alibaba released Qwen 3.6-Plus on April 2, 2026. It is a closed-source model available only via API (OpenRouter, Alibaba Cloud). Key numbers:

MetricQwen 3.6-PlusClaude Opus 4.5GPT-5.4
SWE-bench Verified78.8%80.9%
Terminal-Bench 2.061.659.3
Context Window1M tokens200K128K
Throughput158 tok/s93.5 tok/s76 tok/s
What this means for local users: Qwen 3.6-Plus is API-only, so you cannot run it via Ollama. The open-weight Qwen 3.5 Medium series remains your best option for local inference. However, 3.6-Plus shows where the architecture is heading — expect open-weight successors later in 2026.

Qwen3-Coder-Next — Best Local Coding Model (March 2026)

Qwen3-Coder-Next is a specialized 80B MoE model with only 3B active parameters, designed specifically for agentic coding. It scores 70.6-71.3% on SWE-bench depending on the scaffold.
SpecQwen3-Coder-NextQwen3.5-35B-A3B
Total Params80B35B
Active Params3B3B
Context256K128K
RAM (Q4)~46 GB~20 GB
Best ForCoding agentsGeneral reasoning
Mac requirement: You need a Mac Studio with 64GB+ unified memory to run Coder-Next comfortably. For MacBook Pro 24-32GB, the 35B-A3B remains the better all-rounder.
# Run Qwen3-Coder-Next locally

ollama run qwen3-coder-next

Qwen 3.5 Small Models — Tiny but Powerful (March 2026)

Alibaba released four small models on March 2, 2026: 0.8B, 2B, 4B, and 9B. All are natively multimodal (text + image + video), Apache 2.0 licensed, and support 256K context.

ModelSize (Q4)GPQA DiamondMMMU-ProBest For
Qwen3.5-9B6.6 GB81.7%70.1%General + vision
Qwen3.5-4B3.4 GBEdge devices
Qwen3.5-2B2.7 GBLightweight tasks
Qwen3.5-0.8B1.0 GBEmbedded / mobile

The 9B model is remarkable: it matches or surpasses GPT-OSS-120B (a model 13x its size) on GPQA Diamond and HMMT benchmarks. For MacBook Air users with 8GB RAM, the 4B is now a strong recommendation.

# Small models — run on any Mac

ollama run qwen3.5:9b

ollama run qwen3.5:4b

ollama run qwen3.5:2b

ollama run qwen3.5:0.8b

Ollama 0.19 + MLX — Up to 93% Faster on Apple Silicon (March 31, 2026)

This is the biggest quality-of-life update for Mac users. Ollama 0.19 now uses Apple's MLX framework natively, delivering:

  • 57% faster prefill (time to first token)
  • 93% faster decode (tokens per second)
  • Smarter unified memory management
  • M5/M5 Pro/M5 Max GPU Neural Accelerator support
Practical impact: The Qwen3.5-35B-A3B that ran at ~45 tok/s on M4 now runs closer to 70-80 tok/s with MLX enabled. This makes interactive coding and chat feel nearly instant.
Note: MLX support in Ollama 0.19 is a preview release. Currently optimized for Qwen3.5 models, with broader model support coming soon.

Updated Qwen Ecosystem Map (April 2026)

ModelTypeParamsActiveRAM (Q4)ContextLocal?
Qwen3.5-0.8BDense0.8B0.8B1 GB256KYes
Qwen3.5-2BDense2B2B2.7 GB256KYes
Qwen3.5-4BDense4B4B3.4 GB256KYes
Qwen3.5-9BDense9B9B6.6 GB256KYes
Qwen3.5-27BDense27B27B16 GB128KYes
Qwen3.5-35B-A3BMoE35B3B20 GB128KYes
Qwen3.5-FlashMoE35B22 GB1MYes
Qwen3.5-122B-A10BMoE122B10B72 GB128KYes
Qwen3-Coder-NextMoE80B3B46 GB256KYes
Qwen3.6-PlusHybrid1MAPI only

Use Cases

Qwen3.5-9B — Best for 8-16GB Macs

  • Natively multimodal (text, image, video)
  • Matches models 13x its size on reasoning benchmarks
  • Ideal for: Vision tasks, general chat, lightweight coding, edge deployment

Qwen3.5-Flash — Production Agents

  • 1M context by default (analyze entire documents)
  • Built-in tools (calculator, search, code execution)
  • Ideal for: Autonomous agents, advanced RAG, long document analysis

Qwen3.5-35B-A3B — The Sweet Spot

  • Near-frontier quality on a standard MacBook
  • ~75 tok/s with Ollama 0.19 MLX on M4 Pro
  • Ideal for: Coding, complex reasoning, intelligent chatbots

Qwen3-Coder-Next — Dedicated Coding Agent

  • 80B MoE with 3B active, 256K context
  • 70.6-71.3% on SWE-bench (close to frontier closed models)
  • Ideal for: Agentic coding, repo-level tasks, multi-file refactoring

Qwen3.5-122B-A10B — Local Frontier

  • GPT-4 / Claude 3.5 level quality
  • Requires Mac Studio Ultra
  • Ideal for: Enterprise, research, critical tasks

How to Test Them

Via Ollama (Local)

All Qwen 3.5 models are available on Ollama (58 tags, 4.6M+ downloads). Update to Ollama 0.19+ for MLX acceleration on Apple Silicon.

# Small models (any Mac)

ollama run qwen3.5:9b # 6.6 GB — best small model

ollama run qwen3.5:4b # 3.4 GB — great for 8GB Macs

# Medium models (24GB+ Macs)

ollama run qwen3.5:35b-a3b # 20 GB — best perf/RAM ratio

ollama run qwen3.5:27b # 16 GB — dense, reliable for coding

ollama run qwen3.5:flash # 22 GB — 1M context window

# Large models (Mac Studio)

ollama run qwen3.5:122b-a10b # 72 GB — frontier quality

ollama run qwen3-coder-next # 46 GB — best coding agent

# Tiny models (edge / embedded)

ollama run qwen3.5:2b # 2.7 GB

ollama run qwen3.5:0.8b # 1.0 GB

Tip: On Ollama 0.19+, enable MLX for up to 93% faster decode speeds on Apple Silicon. Check with ollama --version and update if needed.

Via Cloud API

Our modelfit.io Recommendation (April 2026)

After testing with Ollama 0.19 MLX, here are our updated recommendations:

MacBook Air M2/M3/M4 8GB

Qwen3.5-4B (3.4 GB, ~90 tok/s) — Multimodal, surprisingly capable

MacBook Air M3/M4 16GB

Qwen3.5-9B (6.6 GB, ~120 tok/s) — Beats models 13x its size

MacBook Air M3/M4 24GB

Qwen3.5-27B (16 GB, ~85 tok/s) — Dense, reliable for coding

MacBook Pro M4 24-32GB

Qwen3.5-35B-A3B (20 GB, ~75 tok/s) ⭐ Best all-rounder

Mac Studio M4 64-96GB

Qwen3-Coder-Next (46 GB, ~50 tok/s) — Best local coding agent

Mac Studio M4 Ultra 128GB+

Qwen3.5-122B-A10B (72 GB, ~45 tok/s) — Frontier quality

Conclusion

Qwen 3.5 marked a turning point in February 2026, and the ecosystem has only gotten stronger since. The era of "bigger just because" models is over. Well-designed MoE architecture delivers frontier performance with 7x better efficiency.

As of April 2026, the Qwen stack covers every Mac: the 4B for 8GB MacBook Airs, the 35B-A3B for 24GB MacBook Pros, and Coder-Next for Mac Studio coding workstations. Ollama 0.19 with MLX nearly doubles inference speeds, making all of these models feel faster than ever.

For modelfit.io users: The 35B-A3B remains our #1 recommendation for MacBook Pro 24GB. For dedicated coding, add Qwen3-Coder-Next if you have the RAM. And if you are on a budget Mac, the 9B punches far above its weight.

Related: Compare Qwen 3.5 with DeepSeek-V3 in our head-to-head comparison, see the latest Qwen 3.5 Small models, or check MacBook Pro and Mac Studio recommendations.

Frequently Asked Questions

How does Qwen 3.5-35B-A3B beat models 7x its size?

The MoE (Mixture of Experts) architecture activates only 3B of the 35B total parameters per token. Alibaba focused on smarter expert routing, higher quality training data, and optimized reinforcement learning. The result: 82.1% MMLU with 20GB RAM vs 79.5% for the 235B model needing 140GB.

Can I run Qwen 3.5-35B-A3B on a MacBook Air?

The model needs 20GB RAM in Q4 quantization. A MacBook Air with 24GB can run it, though you may experience some memory pressure with other apps open. A MacBook Pro with 24GB provides a more comfortable experience with active cooling. For 16GB MacBook Airs, use the 9B model instead.

What is the difference between Qwen 3.5 Flash and 35B-A3B?

Flash offers 1M token context (vs 128K for 35B-A3B), making it ideal for analyzing entire documents and codebases. The 35B-A3B delivers higher quality per token for shorter tasks. Both need similar RAM (~20-22GB). Choose Flash for RAG and long documents, 35B-A3B for coding and reasoning.

How fast is Qwen 3.5-35B-A3B on Apple Silicon after the MLX update?

With Ollama 0.19 and MLX enabled, expect approximately 70-80 tokens per second on MacBook Pro M4 with 32GB. Before the MLX update, speeds were around 45 tok/s. The 93% faster decode speed from Ollama's MLX integration makes interactive coding feel nearly instant. The MoE architecture keeps speed high because only 3B parameters compute per token.

Is Qwen 3.5-122B worth the extra RAM over 35B-A3B?

The 122B model scores higher on quality benchmarks (84.8% vs 82.1% MMLU) but needs 72GB RAM and a Mac Studio Ultra. For most users, the quality difference does not justify the 3.5x RAM increase. The 35B-A3B is the practical choice unless you need frontier-level accuracy.

Should I use Qwen3-Coder-Next or 35B-A3B for coding?

Qwen3-Coder-Next scores 70.6-71.3% on SWE-bench and is purpose-built for agentic coding — multi-file refactoring, repo-level tasks, and autonomous debugging. However, it needs 46GB RAM. If you have a Mac Studio with 64GB+, use Coder-Next for dedicated coding work. On a MacBook Pro with 24-32GB, the 35B-A3B handles coding well alongside general reasoning tasks.

What is Qwen 3.6-Plus and can I run it locally?

Qwen 3.6-Plus is Alibaba's new flagship model released April 2, 2026. It matches Claude Opus 4.5 on Terminal-Bench 2.0 (61.6 vs 59.3) and offers 1M context. However, it is API-only — the weights are not publicly available. For local inference on Mac, the Qwen 3.5 series remains the best open-weight option. Access 3.6-Plus via OpenRouter or Alibaba Cloud.

How does Qwen 3.5-9B compare to larger models?

The 9B model is the standout of the Small series. It matches or surpasses GPT-OSS-120B (13x its size) on GPQA Diamond (81.7 vs 71.5) and HMMT Feb 2025 (83.2 vs 76.7). It runs on any Mac with 8GB+ RAM at over 120 tok/s with MLX. For users who cannot afford 24GB of RAM for the 35B-A3B, the 9B is the clear second choice.

---

Article originally published February 24, 2026. Last updated April 4, 2026 with Qwen 3.6-Plus, Qwen3-Coder-Next, Qwen 3.5 Small series, and Ollama 0.19 MLX speed improvements. Resources:

Have questions? Reach out on X/Twitter