A well-received PSA on r/LocalLLaMA argues that convenience layers such as Ollama and LM Studio can change model behavior enough to distort evaluation. The more durable lesson from the thread is reproducibility: hold templates, stop tokens, sampling, runtime versions, and quantization constant before judging a model.
#llama-cpp
A LocalLLaMA thread highlighted ongoing work to add NVFP4 quantization support to llama.cpp GGUF, pointing to potential memory savings and higher throughput for compatible GPU setups.
A high-engagement LocalLLaMA follow-up benchmark reports that Qwen3.5-35B-A3B runs best on the tested RTX 5080 setup with Q4_K_M quantization, KV q8_0, and --fit without explicit batch flags.
A high-engagement r/LocalLLaMA thread reports strong early results for Qwen3.5-35B-A3B in local agentic coding workflows. The original poster cites 100+ tokens/sec on a single RTX 3090 setup, while comments show mixed reproducibility and emphasize tooling, quantization, and prompt pipeline differences.
A high-signal LocalLLaMA thread points to llama.cpp Discussion #19759, where maintainers say the ggml team is joining Hugging Face while continuing full-time support for ggml and llama.cpp.
A technical r/LocalLLaMA thread pointed to llama.cpp PR #19765, merged on February 20, 2026. The patch unifies parser paths as a stop-gap for Qwen3-Coder-Next issues and adds parallel tool-calling plus JSON schema fixes.
A high-scoring Hacker News thread highlighted announcement #19759 in ggml-org/llama.cpp: the ggml.ai founding team is joining Hugging Face, while maintainers state ggml/llama.cpp will remain open-source and community-driven.
A popular LocalLLaMA post highlights draft PR #19726, where a contributor proposes porting IQ*_K quantization work from ik_llama.cpp into mainline llama.cpp with initial CPU backend support and early KLD checks.
A high-signal r/LocalLLaMA thread tracked the merge of llama.cpp PR #19375 and highlighted practical throughput gains for Qwen3Next models. Both PR benchmarks and community tests suggest meaningful t/s improvements from graph-level copy reduction.
A high-signal r/LocalLLaMA thread tracked the merge of llama.cpp PR #19375 and highlighted practical throughput gains for Qwen3Next models. Both PR benchmarks and community tests suggest meaningful t/s improvements from graph-level copy reduction.