Daniel Vaughan’s Gemma 4 writeup tests whether a local model can function as a real Codex CLI agent, with the answer depending less on benchmark claims than on very specific serving choices. The key lesson is that Apple Silicon required llama.cpp plus `--jinja`, KV-cache quantization, and `web_search = "disabled"`, while a GB10 box worked through Ollama 0.20.5.
#local-llm
RSS FeedA detailed `r/LocalLLaMA` benchmark reports that pairing `Gemma 4 31B` with `Gemma 4 E2B` as a draft model in `llama.cpp` lifted average throughput from `57.17 t/s` to `73.73 t/s`.
A high-scoring LocalLLaMA post argued that merging llama.cpp PR #21534 finally cleared the known Gemma 4 issues in current master. The community focus was not just the fix itself, but the operational details around tokenizer correctness, chat templates, memory flags, and the warning to avoid CUDA 13.2.
A LocalLLaMA post argues that recent llama.cpp fixes justify refreshed Gemma 4 GGUF downloads, especially for users relying on local inference pipelines.
A practical Reddit debugging post argues that a Qwen 3.5 chat-template issue, not the inference engine itself, can invalidate prefix-cache reuse after tool-heavy turns and waste large amounts of compute.
A recent r/LocalLLaMA post presents Qwen3.5 27B as an unusually strong local inference sweet spot. The author reports about 19.7 tokens per second on an RTX A6000 48GB with llama.cpp and a 32K context, while the comments turn into a detailed debate about dense-versus-MoE VRAM economics.
A high-signal r/LocalLLaMA thread is circulating practical Gemma 4 fine-tuning guidance from Unsloth. The post claims Gemma-4-E2B and E4B can be adapted locally with 8GB VRAM, about 1.5x faster training, roughly 60% less VRAM than FA2 setups, and several fixes for early Gemma 4 training and inference bugs.
A LocalLLaMA post with roughly 350 points argues that Gemma 4 26B A3B becomes unusually effective for local coding-agent and tool-calling workflows when paired with the right runtime settings, contrasting it with prompt-caching and function-calling issues the poster saw in other local-model setups.
A recent LocalLLaMA discussion shared results from Mac LLM Bench, an open benchmark workflow for Apple Silicon systems. The most useful takeaway is practical: dense 32B models hit a clear wall on a 32 GB MacBook Air M5, while some MoE models offer a much better latency-to-capability tradeoff.
A high-signal LocalLLaMA post described a port of llama2.c to classic Mac OS that runs Karpathy’s TinyStories 260K model on a stock iMac G3. The project is compelling because most of the work is systems engineering: endianness fixes, memory partition management, and layout debugging on vintage hardware.
A practical HN gist lays out how to run Ollama and Gemma 4 on an Apple Silicon Mac mini, including auto-start, periodic preload, and `OLLAMA_KEEP_ALIVE=-1`. The author says `gemma4:26b` nearly exhausted 24GB unified memory, making the default 8B model a safer operational choice.
A strong r/LocalLLaMA reaction suggests PrismML’s Bonsai launch is landing as more than another compression headline. The discussion combines the company’s end-to-end 1-bit claims with early hands-on reports that the models feel materially more usable than earlier BitNet-style experiments.