$3000: What GPU for my use case? Will my setup work? Dedicated machine instead?

Posted by CrayCJ@reddit | LocalLLaMA | View on Reddit | 21 comments

Hi! My PC has the following specs:

GPU – to be replaced, see below* AMD Radeon RX 5700 XT, 8GB (Powercolor Red Devil)
CPU AMD Ryzen 5 3600X, 3.8 GHz, 6-core
RAM 2x 8GB, 3600 MHz, DDR4 (G.Skill Trident)
Motherboard MSI B450 Gaming Pro Carbon AC
Power supply Corsair HX 750 Platinum
Storage 1TB – Adata SX8200
Operating system Windows 11

My LLM use case: Academic research based on text and creation of teaching materials. I would like for it to be able to handle 50–100 mainly text pdf-files i.e. entire books, understand, search, compare and summarize their content, assess and comment on the content across all files according to specific questions, output quotes incl. exact page numbers. Also, the LLM should help with creating teaching materials based on a different stack of documents, mainly textbook pdfs – teaching materials like worksheets and even dossiers, if possible powerpoint presentations, incl. pictures searched from the internet, ... In that regard, a certain creativity is welcome. No video, audio or picture generation. Occational and limited statistical work and coding. Autonomous use of the computer's UI – like antigravity seems to be able to – would be very helpful, but is probably not possible yet? My LLM knowledge is not quite extensive yet...

My questions:

  1. I cannot estimate what size of LLM or which i need for this usecase, but my *current GPU is not good enough. Claude recommends an LLM size of 70B, at least 34B and thinks a RTX 3090 to be enough and is happy with my other components. I have a budget of about $3000, approx. the price of a RTX 4090/5090 here. What is your recommendation? Any specific GPU model? Do I need a better power supply?

I would also consider AMD, but nvidia seems to be clearly recommended by this community. I don't know if a 4090 or 5090 is overkill or whether there would be other bottle-necks with those...

  1. As this machine is also a hackintosh, that needs an AMD GPU, I'd like to keep using my current GPU in the lower, slower PCIe 2.0 slot (x4) of my motherbord to connect to the monitor while the new GPU is run in the upper PCIe 3.0 slot (x16) for the LLM. Of course, I will run the LLM on Windows only. Is this dual GPU an issue regarding the LLM?

  2. Or, do you recommend me to build a dedicated LLM machine, which hosts the LLM on the network to be accessed? This of course limits my budget for the GPU.

  3. Can you recommend a specific model? In the end of year thread, I've read good things about Qwen3-Coder-30B-A3B:Q4 and Q8.

Many thanks for your help!