Secondary PC options

Posted by UniqueIdentifier00@reddit | LocalLLaMA | View on Reddit | 8 comments

Hey everyone, I’ve been lurking here for while.

I’ve really been enjoying messing around with my 6gb card on my laptop using Gwen 3.5 4B, ollama, and Open WebUI.

One of my friends is gifting me his old PC. It currently has a 3060 8gb in there, and only 8gb of RAM. I’d like to throw Ubuntu on that and use it as a server setup so I can locally access an LLM over my home network. Looks like I can easily do that with my dockerized OpenWebUi setup I’ve been using on my laptop.

My main question is, given my extreme lack of experience in regard to LLM’s, how do I best go about upgrading this PC? The goal is to be able to run 27B - 30B models. I could buy a 3090, but that alone won’t be enough. I could also use the 3060 alongside that to get enough usable VRAM, but I understand there are complications with tying multiple GPU’s together, something related to offloading that I don’t fully understand.

My other consideration is that I could buy 3-4 3060 12gb cards for the price of a single 3090. I don’t know what all it would take power and rack wise to be able to set that up, let alone how to properly use PCI lanes for that.

Next issue is RAM. How much actual RAM do I need to have to be able to use some of there bigger models? I was under the impression that VRAM is what matters, not RAM, to a certain extent.

Thanks for reading and I hope somebody who’s traveled this path can lend a hand. I’m just trying to find the most cost effective way to be able to use some of there larger models. Take care.