Qwen 3.5-35B Uncensored GGUF Models Now Available
1 min readThe Qwen 3.5-35B uncensored variants have been released in optimized GGUF format, providing practitioners with efficient quantization options for local deployment. The "Aggressive" variant preserves the original model weights without personality modifications or refusal filtering, allowing for flexible local inference across diverse use cases.
Community members have been actively benchmarking these models on consumer hardware like RTX 3090 GPUs with various quantization levels (Q3 through Q4 range). These benchmarks demonstrate practical performance characteristics for production deployment, helping users select appropriate quantization levels for their hardware constraints.
The availability of multiple quantization options makes Qwen 3.5-35B accessible to a wider range of hardware configurations, from edge devices to workstation-class systems, while maintaining strong instruction-following and reasoning capabilities.
Source: r/LocalLLaMA · Relevance: 8/10