Questions about revisiting local LLM roleplay.
Posted by newbuildertfb@reddit | LocalLLaMA | View on Reddit | 12 comments
TLDR for those that dojr wanna read below I need a new good free place online to pickup roleplay where should that be and what can I do locally? 9070xt 32gb ram desktop and preferably but I know it not great, 4060 laptop 32gb ram.
First it was GPT/Claude until they remind you before you get very far they are to censored for any real fun. Then Then a few months back (wow September 2025 was closer to a year ago gosh) anyways I tried open router and it was nice for a few weeks then they removed all the DeepSeek or any usable free model (unless they added some I don't know about?) Then as of a few days ago found out Ollama has good DeepSeek but its also taken down now (I think nobody knows what is going on?)
I don't want to pay especially when its a monthly that sounds more sad then I got good GPU but my roleplays have been so fun...I want to pick them back up. What hardware do I need? When open router removed DeepSeek I tried local LLM (9070xt I didn't biy the right hardware for this but got that card not just for that at launch and 4060 laptop) and it could not do the roleplay I wanted to do but idk with advancements, maybe things change? What can it run, how well will it do and if I copy over old chat to new place how close to old chat quality I gonna get? I was doing anime fandom roleplays.
BitGreen1270@reddit
If I'm reading you right, you want a local llm that can run on a 9070xt or 4060 but also do uncensored roleplay? Should be possible with gemma4. You can see my previous comment on another thread to get gemma4-26B working and pass in the jailbreak system prompt (which you can get by searching on this sub). I had it work for me 8 times out of 10.
Equivalent_Bit_461@reddit
Bless your soul for that jailbreak
newbuildertfb@reddit (OP)
I wanna find an online replacement for my really large I mashed fairy tail and xenoblade into one anime world and been doing way to much roleplay...that doesn't cost money and anything I need uncensored maybe an nsfw smaller thing I want to do on local if I can. I don't know what models to try though. Will it be any better then 6th ish (give or take a month) months ago when I last tried because then no it won't do the kinda roleplay I want well enough because be my GF stuff is just to boring.
Formal-Exam-8767@reddit
For best local experience SillyTavern coupled with an ERP finetune from TheDrummer that fits your GPU card's VRAM (you can go over but prefill will suffer).
Miriel_z@reddit
HuggingFace, SillyTavern. There are pretty good RP models out there, wild and uncensored.
newbuildertfb@reddit (OP)
Obviously some models are better at rp then others even if it is slightly less parameters by a billion or something but I don't know enough to know what to try other then how many B won't fit or run well enough on my hardware.
What models do I try and will I have a noticeably better experience then when I tried before back in September October? I know Google released their optimization paper thing and its just been a bit now? If it won't be noticeably better then it isn't worth it.
Its not just oh yeah it can play world for me of my GF, it also has to get anime facts right becasw nothing takes me out of it more then having to refresh the message 1-3 extra times because it misses a fact or got x wrong and then I gotta correct it or ESPECIALLY the taking over as me problem I had before. So it gotta be good at OK I know this anime here is a million (not literally) correct facta about it and it can roleplay well.
overand@reddit
HEad over to r/SillyTavernAI for this. they have a weekly thread discussing which models to try, but I'd start with Cydonia-24b-v4.3 and one of the Gemma-4 models (Start with the regular ones, not the uncensored ones - you'll get surprisingly far with a "jailbreak" prompt; I didn't believe it until I tried it.)
123vovochen@reddit
Hey MAYBE, at THIS point, it is time for you to real life, because you are obviously in too deep in what will turn out to be a terribly lonely addiction.
Miriel_z@reddit
Try Wingless Imp, Impish Nemo. These are from my memory. There might be much better variants now though.
newbuildertfb@reddit (OP)
Will do, is it just a roleplay one or is it gonna pass my summarize my fandom show and let me see if you get enough details right to trust doing an in universe test as well?
Also I'm willing good (if it is actually really good again rip Ollama worked out so nice) could if it is free so anywhere I should check for that?
Miriel_z@reddit
You can find quantized models and safetensors themselves. Up to you where you want to run them. I use my own chat/code, gives me more flexibility.
newbuildertfb@reddit (OP)
Ideally I'd find another we have DeepSeek 3.1 or v3 I can use. I think Ollama had a no logs policy idk they were nice but Ollama and open router got rid of that so for cloud I don't know where to go? Doesn't have to be DeepSeek just need good uncensored options.
As for local I'll try some models and take suggestions see what or how well they run.