[Build Help] Best RP models and frontends for 4090 (24GB VRAM) / 64GB RAM? (No SillyTavern)

Posted by Ok_Cartographer_809@reddit | LocalLLaMA | View on Reddit | 8 comments

Hi everyone,

I'm looking for some recommendations to level up my local RP experience. My current setup is a Windows machine with an i7-14700K, 64GB DDR5 RAM, and an RTX 4090 (24GB VRAM).

I am currently using LM Studio, which I like for its ease of use. However, I’m looking for a frontend that is more specialized for Roleplay—specifically something with robust support for Character Cards and Memory/Lorebook features—without going down the SillyTavern rabbit hole.

For models, since I have 24GB of VRAM and plenty of system RAM, what are the current "S-Tier" recommendations for high-quality, creative RP in 2026? I’m interested in models that:

  1. Excel at nuanced prose and avoiding "GPT-isms."

  2. Can handle long-context roleplay without losing character consistency.

  3. Fit well within my hardware (I'm open to GGUF or EXL2).

Questions:

  1. Is there a frontend that bridges the gap between LM Studio's simplicity and SillyTavern's features? (e.g., Faraday/AnythingLLM/etc.)

  2. Which 30B-70B models are currently the favorites for immersive storytelling on a single 4090?

Thanks for the help