Introducing Wayfarer: a brutally challenging roleplay model trained to let you fail and die.

Posted by Nick_AIDungeon@reddit | LocalLLaMA | View on Reddit | 87 comments

One frustration we’ve heard from many AI Dungeon players is that AI models are too nice, never letting them fail or die. So we decided to fix that. We trained a model we call Wayfarer where adventures are much more challenging with failure and death happening frequently.

We released it on AI Dungeon several weeks ago and players loved it, so we’ve decided to open source the model for anyone to experience unforgivingly brutal AI adventures!

Would love to hear your feedback as we plan to continue to improve and open source similar models.

[https://huggingface.co/LatitudeGames/Wayfarer-12B](https://huggingface.co/LatitudeGames/Wayfarer-12B)