Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

2021-EMNLP-Improving Neural Machine Translation by Bidirectional Training #245

Open
thangk opened this issue Jun 27, 2024 · 0 comments
Open
Assignees
Labels
literature-review Summary of the paper related to the work

Comments

@thangk
Copy link
Collaborator

thangk commented Jun 27, 2024

Link: arXiv

Main problem

Improving the performance of NMT model translation is hard. Even when there are high-performing models, they're often configured with much more complicated settings. That limits its applicability to a broader range of tasks. Thus, there needs to be a solution that's simple and yet can be applied to more tasks.

Proposed method

The author proposes a method that involves pre-training the data that goes from source->target then to source+target->target+source. This doubles the training data. The model is then tuned for the normal direction of source->target direction.

My Summary

According to the paper's result, this proposed method significantly improved the performance than a regular one-directional pre-training approach. The author observes the best BLEU scores in IWSL21 in low-resource tasks. The author claims this is a better and simple bilingual code-switching approach and also improves bilingual alignment quality. There are more testing needed to be done, such as if this bidirectional pre-training can be applied to previous NMT systems.

Datasets

IWSLT21 EN-DE
WMT16 EN-RO
WMT19 EN-GU
IWSLT21 EN-SW
WMT14 EN-DE
WMT19 EN-DE
WMT17 ZH-EN
WAT17 JA-EN
@thangk thangk self-assigned this Jun 27, 2024
@thangk thangk added the literature-review Summary of the paper related to the work label Jun 27, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
literature-review Summary of the paper related to the work
Projects
None yet
Development

No branches or pull requests

1 participant