Can be used for machine translation, speech translation, language modeling and dialogue supporting a number of popular pre-trained models.
Code repository for paper: What do compressed multilingual machine translation models forget?
Covers more than 10K language pairs, achieves competitive results with M2M-100 while being much smaller and faster.
Publications concern efficient inference, continual learning, unsupervised NMT and domain adaptation.