Comprehensive MoE Backend Benchmarks for Qwen3.5-397B: Real Numbers vs Hype

1 min read
r/LocalLLaMAcommunity

A comprehensive 8+ hour benchmark study tested every major MoE inference backend for Qwen3.5-397B on workstation-grade hardware (4x RTX PRO 6000). The findings are sobering: actual sustained decode performance reaches only 50.5 tok/s, despite claims of 130+ tok/s floating around online.

The investigation reveals that Nvidia's own CUTLASS kernels have performance issues on their own workstation GPUs, suggesting broader problems in the optimization stack. This is critical for practitioners evaluating large MoE models for local deployment—claims about theoretical throughput often don't survive real-world testing. The benchmark methodology and findings provide a reality check on what's actually achievable with current hardware and software stacks.

For anyone considering 300B+ parameter models, this data-driven analysis cuts through marketing claims and provides honest expectations about performance. Understanding actual sustained throughput, not peak numbers, is essential for production planning and hardware procurement decisions.


Source: r/LocalLLaMA · Relevance: 8/10