MiniCPM4: Ultra-Efficient LLMs on End Devices
Posted by adefa@reddit | LocalLLaMA | View on Reddit | 7 comments
Randomly saw this -- no models yet.
Posted by adefa@reddit | LocalLLaMA | View on Reddit | 7 comments
Randomly saw this -- no models yet.
Iory1998@reddit
EDIT:
There is an 8B model right now.
Pedalnomica@reddit
There are models now... Quite a few. Kind of overwhelming TBH!
sanobawitch@reddit
Do we have any QAT trainer as open source code?
Their 3.3.1 Efficient Quantization-Aware Training in the technical report doesn't tell much.
secopsml@reddit
How they will name their models in 2 years if those are already
ultra
?terminoid_@reddit
no embedding models?
minpeter2@reddit
Looks good, I'm looking forward to it!!
unknownwarriorofmars@reddit
openbmb my beloved.