What can I run in macbook pro m4 16gb
Posted by distan_to-reality_66@reddit | LocalLLaMA | View on Reddit | 5 comments
Sand as the title, what models can I run
Posted by distan_to-reality_66@reddit | LocalLLaMA | View on Reddit | 5 comments
Sand as the title, what models can I run
ttkciar@reddit
Please respond to this thread in the model recommendation megathread only! https://old.reddit.com/r/LocalLLaMA/comments/1sknx6n/best_local_llms_apr_2026/
InteractionSmall6778@reddit
16GB unified memory on Apple Silicon gives you more usable RAM for inference than an equivalent PC setup.
You can comfortably run Qwen3 14B at Q4_K_M, Gemma 3 12B, or Phi-4 14B Q4 at solid speeds with llama.cpp or Ollama. For vision tasks, Llama 3.2 11B Vision fits well within your limit.
If you want to leave headroom for other apps, Llama 3.2 3B or Gemma 3 4B are blazingly fast options too.
Alexandratang@reddit
All these models are outdated because you are a bot. New ones have released post your training data cut off, which seems to have been in mid 2025.
distan_to-reality_66@reddit (OP)
I mainly develop agents to use as bounce board for my writing (mostly screenplay and scripts)
SpringBeginning8897@reddit
Es una herramienta basada en terminal (escrita en Rust) pensada para entusiastas de la IA local.