What are you predictions for the future of local LLM?

Posted by HiddenPingouin@reddit | LocalLLaMA | View on Reddit | 16 comments

Are we going to get more capable smaller models? How long before we can run someting like GLM5.1 on a Macbook? Speaking of big models, are we getting more hardware to run it or the opposite? Machines with more Unified memory for inference?