Gemma 4 26B-A4B GGUF Benchmarks

Posted by danielhanchen@reddit | LocalLLaMA | View on Reddit | 106 comments

Gemma 4 26B-A4B GGUF Benchmarks

Hey r/LocalLLaMA we conducted KL Divergence benchmarks for Gemma 4 26B-A4B GGUFs across providers to help you pick the best quant.

For HQ versions of the graphs as Reddit mobile compresses it. See: Gemma 4 Benchmarks and Qwen3.6 Benchmarks

We also updated our MLX quants to be more dynamic with better layering selection (there are limitations due to MLX): See here

MLX Metrics UD-4bit (Old) UD-4bit (New) MLX 4.4bit MSQ
Perplexity 4.772 4.766 4.864
Mean KLD 0.0177 0.0163 0.0878
99.9% KLD 0.8901 0.8398 2.9597
Disk Sze 21.4 GB 21.6 GB 21.2 GB

Gemma 4 GGUFs: https://huggingface.co/unsloth/gemma-4-26B-A4B-it-GGUF

Qwen3.6 GGUFs: https://huggingface.co/unsloth/Qwen3.6-35B-A3B-GGUF