How I use Gemma 3 to help me reply my texts

Posted by sean01-eth@reddit | LocalLLaMA | View on Reddit | 28 comments

Ever since there're code completions, I wish I could have something similar when texting people. Now there's finally a decent method for that.

The app works on any endpoint that's OpenAI compatible. Once you set it up, it gives you texting completions right inside WhatsApp, Signal, and some other texting apps.

I tested it with Gemma 3 4B running on my AMD Ryzen 4700u laptop. The results come out slow, but the quality is totally acceptable (the video is trimmed, but the suggestions come from Gemma 3 4B). I can imagine if you have a powerful setup, you can get these texting suggestions with a fully local setup!

Here's a brief guide to make this work with ollama:

My laptop isn't powerful enough, so for daily use, I use Gemini 2.0 Flash, just change the URL, API Key, and model name.

Let me know how's your experience with it!