Hi all,
In the age of decoder-only LLMs, I'll like to ask if there's any competitive encoder-decoder architectures that are known to scale well for multilingual seq2seq tasks?
There's these that reported state-of-the-art NLI scores but they were not known to be multilingual
Other than the above, are there any competitive encoder-decoder architectures that are known to scale well for multilingual seq2seq tasks?
Thank you in advance for the pointers!
Regards,
Liling