Back to home
Qwen3-Coder
2 articles tagged with this topic
llama.cppGLM-4.7
Best Local LLM for Agentic Coding on a Single RTX 4090
A 4090 owner benchmarks GLM-4.7, Nemotron-30B, and Qwen3-Coder for local agentic coding via llama.cpp.
Apr 61 min read
Qwen3-Coderllama.cpp
Run Qwen3-Coder 80B Locally at 54GB With Apex Quantization
A community GGUF quantization shrinks Qwen3-Coder 80B to 54.1GB, making fast local coding inference practical.
Apr 52 min read