Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

how to use multi-gpus to finetune large datasets like ReCoRD? #1338

Open
runzeer opened this issue Sep 2, 2021 · 3 comments
Open

how to use multi-gpus to finetune large datasets like ReCoRD? #1338

runzeer opened this issue Sep 2, 2021 · 3 comments

Comments

@runzeer
Copy link

runzeer commented Sep 2, 2021

No description provided.

@zphang
Copy link
Collaborator

zphang commented Sep 6, 2021

The main runscript should automatically detect the number of GPUs available and run with DataParallel. Are you encountering any issues doing so?

@runzeer
Copy link
Author

runzeer commented Sep 6, 2021

when using the DataParallel, the memory cost in each GPU is not equal. So if convenient, could you change the mode to DDP?

@zphang
Copy link
Collaborator

zphang commented Sep 9, 2021

Yep, we are aware that alternative multi-GPU methods like DDP have some advantages over DataParallel, but switching to DDP will introduce non-trivial complexity, and we do not currently have plans to incorporate it, although we are not ruling it out for the future.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants