My thought on Qwen and Gemma

Posted by Internal-Thanks8812@reddit | LocalLLaMA | View on Reddit | 34 comments

This spring is really hot since the localLLM giant, both Qwen and Gemma released major models.
I'm really excited with those release and happy with their capability.
Both are real hero for local LLM, although I have feeling they have different strength.
For the background, I use them with text review, grammar check in human/social science field and some coding with python(mostly light data analysis stuff), web app(js, ts), general stuff.
I use 27/31B dense and 35/26B Moe, haven't much tried with smaller models.

Qwen
Strength

Weakness

Gemma

Strength

Weakness

Bias

Both has bias in different way/direction, especially politics/cultural topic. Since I believe real "neutral" model is impossible in general, I would always keep it in my mind. But I feel Qwen got more toward to neutral since 3.5(before it was much biased in my opinion), similar neutrality to Gemma.

They still hallucinate occasionally and sometimes dumm, but I think it is also good for me since I still need to use my brain/hand to cover it to avoid got Alzheimer.

Both are open weight, I continue use them by case.
My usage is not so much heavy, so I may miss something and this is just my opinion/feelings.
What is your thought? I'm curious.