Looking for a LLM that is close to gpt 4 for writing or RP

Posted by Intrepid-Biscotti912@reddit | LocalLLaMA | View on Reddit | 7 comments

Hey everyone,

Quick question: with 288GB of VRAM, what kind of models could I realistically run? I won’t go into all the hardware details, but it’s a Threadripper setup with 256GB of system RAM.

I know it might sound like a basic question, but the biggest I’ve run locally so far was a 13B model using a 3080 and a 4060 Ti. I’m still pretty new to running local models only tried a couple so far and I’m just looking for something that works well as a solid all-around model, or maybe a few I can switch between depending on what I’m doing.