[New paper] Cross-Domain Content Generation with Domain-Specific Small Language Models
Posted by ankitm1@reddit | LocalLLaMA | View on Reddit | 1 comments
ankitm1@reddit (OP)
Context: cos i could not do on the link.
If I understand this correctly, this paper enables a model to expand it's training to unseen tokens, without having to train from scratch, and without forgetting previous knowledge. (no catastrophic forgetting). Any thoughts on this approach?