You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I don't think there would be a big problem with your suggestion, except that a very small batch could theoretically yield a noisy gradient. Please try and let me know if you run into problems.
Currently, an error occurs when the epoch_size argument is not an integer multiple of batchsize.
attention-learn-to-route/options.py
Line 86 in c66da2c
I assume this is to ensure that there is no batch with less than the batchsize amount of data samples.
Is there any reason not to use the droplast parameter in DataLoaders to throw away such batches instead of generating an error?
The text was updated successfully, but these errors were encountered: