Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Machine Translation Using Transformers namely T5 and XLM-RoBERTa #17

Open
nevasini opened this issue Apr 4, 2023 · 1 comment
Open

Comments

@nevasini
Copy link

nevasini commented Apr 4, 2023

I plan to implement 2 transformer models, and finally choose the one which performs the best on evaluation.
T5 (Text-to-Text Transfer Transformer) is a powerful and flexible Transformer-based language model that can be fine-tuned for a variety of natural language processing tasks, including machine translation. Unlike traditional machine translation models that are designed specifically for translation, T5 can be fine-tuned on a diverse set of text-to-text tasks, including translation that we desire to work on.
Whereas XLM-RoBERTa (Cross-lingual Language Model RoBERTa) https://arxiv.org/pdf/1911.02116.pdf is a state-of-the-art multilingual language model developed by Facebook AI. It is based on the RoBERTa architecture, which is a variant of the BERT model that has been pre-trained on a large corpus of unlabeled text data using a masked language modeling objective.
The key innovation of XLM-RoBERTa is its ability to perform cross-lingual language modeling, which means it can learn representations of multiple languages simultaneously

@vikasgrewal16
Copy link

can i get to know something about this issue ? @nevasini

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants