Someone recently ran an LLM on a 1998 model iMac with 32 MB of RAM. How did you push this boundary and found an usable LLM that also scales well on CPU?

Posted by last_llm_standing@reddit | LocalLLaMA | View on Reddit | 13 comments

Which SLM has proven to give the most throughput, does decent reasoning, and can run fast on a 16/32GB RAM machine based on your experiments?