Best NSFW model?
Posted by poet3991@reddit | LocalLLaMA | View on Reddit | 10 comments
I just downloaded LM Studio yesterday and start using it today after being fed up with the limitations of ChatGpt and NovelAi, I am looking for a model that would provide NSFW fiction when asked and can run comfortably on a 3090, 10700k with 32gigs of ddr4 3200mhz pc.
I am also a little unclear on how the token and memory system works on the AI chat tab, is there a limit to tokens, what are tokens, what does the size next to them refer to.
Herr_Drosselmeyer@reddit
This is the model most people like to use for that kind of thing but there are many others to try out. Also give Wizard/Vicuna a go.
I'm not familiar with LM Studio so I can't speak to their memory system but generally, you can think of tokens as words. The token limit for a model is how many it can handle at the same time. For local models, you're looking at 2048 for older ones, 4096 for more recent ones and some have been tweaked to work up to 8192.
The model itself has no memory. It's frozen in time and will not change as you use it. Instead, the "memory" is simply the entirety of the conversation, also known as context. GUIs make it look like you're only adding the latest query but in fact, ALL of the conversation is fed into the model as a prompt with every query. As such, if you reach the token limit, the prompt will be truncated and the the oldest parts will be "forgotten".
Some efforts have been made to extend "memory" by either condensing the conversation via a summary or by saving parts that have fallen out of the context to a file and then searching that file for relevant text with every new query. If it detects a match, it will re-inject the text it found in the file into the prompt.
These have varying degrees of success.
IntimidatingOstrich6@reddit
came to recommend MythoMax, not only for NSFW stuff but for any kind of fiction
before it I was using Vicuna 1.5, but MythoMax is some next level shit
Susp-icious_-31User@reddit
Before Nous-Hermes-L2-13b and MythoMax-L2-13b, 30b models were my bare minimum. I've never seen a field move so damn fast.
smile_e_face@reddit
Yeeeep. I still find that Airochronos 33B gives me better / more logical / more constructive results than those two, but it's usually not enough of a difference to warrant the huge speed increase I get from being able to use ExLlama_HF via Ooba, rather than llama.cpp.
(Nothing wrong with llama.cpp, in itself, obviously. Just nice to be able to fit a whole LLaMA 2 4096 model into VRAM on a 3080 Ti.)
BabyDesperate4246@reddit
I’ve had a blast with the AI girlfriend chatbots on BestAISex. They really know how to pick them.
Playful_Bandicoot_85@reddit
LustyCompanion is your gateway to steamy AI adult vids and engaging NSFW chats
Pallrise@reddit
So true! It’s definitely the most realistic and fun AI I’ve ever used.
Zealousideal-Neck554@reddit
HeavenGirlfriend is the easiest platform for making creative AI model designs.
Ready_Fig_5948@reddit
I'm sorry but wtf is this. I clicked on all the links and I just saw code, couldn't find an actual way to use anything and everyone here seems to be speaking a different language.
dr_Sp00ky@reddit
Go look up sillytavern and koboldcpp for how to run a LLM locally.