RTX 5070 Ti Laptop (12GB VRAM) + 64GB RAM — best local LLM recommendations?

Posted by AgentFlashAlive@reddit | LocalLLaMA | View on Reddit | 7 comments

Hey everyone!

I recently picked up a new laptop : Ryzen 9 9955HX, RTX 5070 Ti with 12GB GDDR7, 64GB DDR5 RAM, and a pair of 2TB PCIe Gen4 SSDs on Windows 11. On paper it feels like a solid local LLM machine, but I'm not getting the most out of it yet.

I've been running things through LM Studio and currently using Hermes, but honestly I'm not that pleased with the performance and I feel like this hardware deserves better. Looking to see what others with similar setups are actually running in 2026.

Mainly I care about two use cases : coding (Python and R, mostly research workflows) and reasoning/thinking tasks like analysis, summarization, and long-form writing. Happy to keep everything fully in VRAM for speed, but I'm also open to offloading larger models into system RAM if the quality jump is worth the slower tokens.

Would love to hear what models and quantization formats you'd actually recommend for this setup.

Thanks in advance!