Anyone running local LLM coding setups on 24GB VRAM laptops? Looking for real-world experiences

Posted by AmazinglyNatural6545@reddit | LocalLLaMA | View on Reddit | 31 comments

Hi everyone

I’m wondering if anyone has real day-to-day experience with local LLM coding on 24GB VRAM? And how do you use it? Cline/Continue in VScode?

Here’s the situation: I’ve been using Claude Code, but it’s getting pretty expensive. The basic plan recently got nerfed — now you only get a few hours of work time before you have to wait for your resources to reset. So I’m looking into local alternatives, even if they’re not as advanced. That’s totally fine — I’m already into local AI stuff, so I am a bit familiar with what to expect.

Right now I’ve got a laptop with an RTX 4080 (12GB VRAM). It’s fine for most AI tasks I run, but not great for coding with LLMs.

For context:

I’m thinking about getting a 5090 laptop, but that thing’s insanely expensive, so I’d love to hear some thoughts or real experiences from people who actually run heavy local AI workloads on laptops.

Thanks! 🙏