Best way to build a 4× RTX 3090 AI server (with future upgrade to 8 GPUs)?

Posted by Lazy_Independent_541@reddit | LocalLLaMA | View on Reddit | 24 comments

I'm planning to build a local AI workstation/server and would appreciate advice from people who have already done multi-GPU setups.

My current idea is to start with 4× RTX 3090 (24GB each) and possibly scale to 8× GPUs later if the setup proves useful.

My main workloads will be:

Coding LLMs for an agentic development setup

Running open-source coding models locally (DeepSeek, CodeLlama, etc.)

Using them with Claude Code–style workflows / coding agents

Image and video generation

Running ComfyUI workflows

Stable Diffusion / video models / multi-GPU inference if possible

Questions

  1. Hardware platformWhat is the best platform for this type of build?

Options I’m considering:

Threadripper / Threadripper Pro

AMD EPYC

Intel Xeon

My goal is to start with 4 GPUs but keep the option to scale to 8 GPUs later without rebuilding everything.

  1. Motherboard recommendationsWhat boards work well for multi-GPU setups like this?

Things I’m trying to avoid:

PCIe lane bottlenecks

GPUs throttling due to slot bandwidth

Compatibility issues with risers

  1. Is 8× 3090 still worth it in 2026?

Since the 3090 is an older card now, I'm wondering:

Is it still a good investment for local AI servers?

What bottlenecks would I face with an 8×3090 system?

Possible concerns:

PCIe bandwidth

power consumption

NVLink usefulness

framework support for multi-GPU inference

  1. Real-world experiences

If you’re running 4× or 8× 3090 setups, I’d love to know:

what CPU / motherboard you used

how you handled power and cooling

whether you ran into scaling limitations

Goal

Ultimately I want a local AI server that can:

run strong coding models for agentic software development

run heavy ComfyUI image/video workflows

remain expandable for the next 2–3 years

Any build advice or lessons learned would be hugely appreciated.