Trained a chess LLM locally that beats GPT-5 (technically)

Posted by KingGongzilla@reddit | LocalLLaMA | View on Reddit | 35 comments

Hi everyone,

Over the past week I worked on a project training an LLM from scratch to play chess. The result is a language model that can play chess and generates legal moves almost 100% of the time completing about 96% of games without any illegal moves. For comparison, GPT-5 produces illegal moves in every game I tested, usually within 6-10 moves.

I’ve trained two versions so far:

The models can occasionally beat Stockfish at ELO levels between 1500-2500, though I’m still running more evaluations and will update the results as I go.

If you want to try training yourself or build on it this is the Github repo for training: https://github.com/kinggongzilla/chess-bot-3000

vRAM requirements for training locally are \~12GB and \~22GB for the 100m and 250m modle respectively. So this can definitely be done on an RTX 3090 or similar.

Full disclosure: the only reason it “beats” GPT-5 is because GPT-5 keeps making illegal moves. Still, it’s been a fun experiment in training a specialized LLM locally, and there are definitely a lot of things one could do to improve the model further. Better data curation etc etc..

Let me know if you try it out or have any feedback!