Llama.cpp Adds True Reasoning Budget Support

1 min read
r/LocalLLaMAcommunity

Llama.cpp has shipped a major feature that the community has been requesting: true reasoning budget support. Previously, the --reasoning-budget parameter was essentially non-functional, serving only to disable thinking entirely. Now users have granular control over how many thinking tokens the model allocates during inference.

This is critical for local deployment because reasoning models like o1 and Qwen3.5 can generate substantial internal thinking tokens that increase latency and VRAM usage. With proper budget control, practitioners can balance response quality against computational cost, making these powerful models viable on resource-constrained hardware. The feature enables optimization strategies like early stopping when sufficient reasoning depth is achieved.

For anyone running reasoning models locally, this update significantly improves cost-performance tradeoffs and makes the inference process more predictable and controllable.


Source: r/LocalLLaMA · Relevance: 9/10