Anyone else experiencing heavy hallucinations with MiMo-V2.5 (310B) quantized version?

Posted by Shoddy_Bed3240@reddit | LocalLLaMA | View on Reddit | 8 comments

Has anyone else run into major issues with MiMo-V2.5 (the 310B total / 15B active MoE model from Xiaomi)?

I tried the UD-Q4_K_XL quant from Unsloth. Use llama.cpp.

It hallucinates really badly, especially on practical tasks. I gave it a list of files to analyze via OpenCode, and it kept messing up filenames and file paths — inventing ones that don't exist, mixing them up, or just confidently wrong about the directory structure.

Has anyone had better luck with other quants (e.g., higher bits like Q5/Q6)?