-
-
Notifications
You must be signed in to change notification settings - Fork 925
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
multipack w batch sampler #795
Conversation
Could this and should this replace the DataLoader? The sampler looks like a simple version of the Multipack DataLoader. |
Yeah. I want to rip out the previous dataloader. Yeah, it's basically the sampler but adapted to subclass the batch sampler and the various fixes needed to use uneven batches |
So Multipack can be implemented as a sampler and then we can use the standard DataLoader from PyTorch? Probably good to make the full adjustment in this PR if we go that way |
|
a369d27
to
b7a516b
Compare
0050a02
to
d7e9af2
Compare
* test batch sampler w varying batch lens * wip * multipack batchsampler wip * wip * fix for prepare data loader to get correct # of steps based on gpues * lint and clean up * calculate len estimate * fix total num steps calc * add options for dataloader_num_workers and dataloader_pin_memory * remove gitbook * support prefetch_factor for dataloader optimization * fix the kwarg
* test batch sampler w varying batch lens * wip * multipack batchsampler wip * wip * fix for prepare data loader to get correct # of steps based on gpues * lint and clean up * calculate len estimate * fix total num steps calc * add options for dataloader_num_workers and dataloader_pin_memory * remove gitbook * support prefetch_factor for dataloader optimization * fix the kwarg
resolves #406