Hi all,
In the age of decoder-only LLMs, I'll like to ask if there's any competitive encoder-decoder architectures that are known to scale well for multilingual seq2seq tasks?
- https://huggingface.co/docs/transformers/en/model_doc/mt5 - https://huggingface.co/facebook/m2m100_418M - https://huggingface.co/google-bert/bert-base-multilingual-cased + https://www.kaggle.com/code/alvations/neural-plasticity-bert2bert-on-wmt14 - https://huggingface.co/Helsinki-NLP/opus-mt-en-mul + https://huggingface.co/Helsinki-NLP/opus-mt-mul-en - https://huggingface.co/docs/transformers/en/model_doc/umt5
There's these that reported state-of-the-art NLI scores but they were not known to be multilingual
- https://huggingface.co/google/ul2 - https://huggingface.co/docs/transformers/en/model_doc/flan-t5 - https://huggingface.co/docs/transformers/en/model_doc/byt5
There's some ideas on doing encoder with mamba https://github.com/state-spaces/mamba/issues/78 but it looks like an open question.
Other than the above, are there any competitive encoder-decoder architectures that are known to scale well for multilingual seq2seq tasks?
Thank you in advance for the pointers!
Regards, Liling