On behalf of our colleague, Laurent Besacier:


hi everyone


i thought this info could be useful for scientists working on NMT for low resource languages
we recently released  SMaLL-100 model, a Shallow Multilingual MT Model for Low-Resource Languages
it is a distilled version of the large 12B MTM-100 model released by Meta

you can find out more here: https://arxiv.org/abs/2210.11621 

but the reason why i want to share this to SIGUL & EAMT community is because we provide models (which are a good pre-trained model to develop MT for low resource language pair) and also a demo platform to access MT for those 10,000 language pairs !

Models: https://huggingface.co/alirezamsh/small100
Online MT demo: https://huggingface.co/spaces/alirezamsh/small100
(still a bit slow because currently running on 2 vCPU - 16GB RAM)

Best regards

Laurent Besacier
Naver Labs Europe
-- 
Claudia Soria
Researcher
Cnr-Istituto di Linguistica Computazionale “Antonio Zampolli”
Via Moruzzi 1
56124 Pisa
Italy

Tel. +39 050 3153166
Skype clausor