Here the R1-Lite-Preview from DeepSeek AI showed its power... WTF!! This is amazing!!
Posted by Inspireyd@reddit | LocalLLaMA | View on Reddit | 9 comments
lordpuddingcup@reddit
I asked it a programming question related to python and apples MLX and it doesnt know what mlx is, felt odd all the other models seem to know it, gap in knowledge dataset i guess
JealousAmoeba@reddit
It's likely a small model, not enough parameters to hold both detailed world knowledge and knowledge about how to reason.
ihaag@reddit
It took 2 shots to answer how many rrrrrrr’s in strawberrrrrrrry but so did Claude latest model, 2 shots asking it ‘are you sure’ I cannot wait for the open weights
Koksny@reddit
That's just base64 decoding with extra steps, is it really something that average model can't do?
EstarriolOfTheEast@reddit
It's impressive. The extra steps matter. No LLM can decode base32, despite some being champs at base64 for example. Open models also tend to be quite bad at decoding ciphers.
Inspireyd@reddit (OP)
For the previous GPT-4o to do this, I had to add some additional information. Before, it would guess a few words and leave the sentence disjointed. Gemini Exp 1114 came the closest.
MLDataScientist@reddit
did you come up with the question or this is an existing question from the internet?
Inspireyd@reddit (OP)
I developed the one in the screenshot myself. I have another basic example in which I simply encrypted a message and it was able to solve it without any problems.
When you go to the internet and ask for things related to the Playfair Cipher, for example, it fails miserably. I don't use o1, someone said on X that o1 can solve encrypted sentences about the Playfair Cipher.
I don't know how to do playfair ciphers, but if GPT-4o is right, the correct answer would be "Yesterday I ate pork chops".
LoadingALIAS@reddit
This just is not a great example.