Heavily quantized Q2 GLM5 vs less quantized Q8 minimax 2.5/Q4 Qwen3.5 397b?

Posted by ImpressiveNet5886@reddit | LocalLLaMA | View on Reddit | 9 comments

How would say the quality compares between heavily quantized versions of higher parameter giant models like GLM-5-UD-IQ2_XXS (241GB size) vs similarly sized but less quantized and fewer parameter models like MiniMax-M2.5-UD-Q8_0 (243GB) or Qwen3.5-397B-A17B-MXFP4_MOE (237GB)?