Back to home
local-LLM
2 articles tagged with this topic
Gemma 4llama.cpp
Fixing Gemma 4 Tool Calls in llama.cpp: Root Causes Explained
Four bugs in llama.cpp's Gemma 4 chat template handling caused tool call results to crash or loop.
Apr 83 min read
Qwen3-Coderllama.cpp
Run Qwen3-Coder 80B Locally at 54GB With Apex Quantization
A community GGUF quantization shrinks Qwen3-Coder 80B to 54.1GB, making fast local coding inference practical.
Apr 52 min read