Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training Becomes Slow When Dataset Length Is Large #155

Open
lltyzzzr opened this issue Dec 7, 2024 · 0 comments
Open

Training Becomes Slow When Dataset Length Is Large #155

lltyzzzr opened this issue Dec 7, 2024 · 0 comments

Comments

@lltyzzzr
Copy link

lltyzzzr commented Dec 7, 2024

Hi!

Thank you for the excellent work on OpenCood! I encountered an issue while training with OpenCood: when the length of the dataset in an epoch exceeds a certain number (I haven't tested it exactly, but I suspect it is between 2500 and 3000), the entire training process becomes significantly slower. To analyze the issue, I tried using a smaller dataset and concatenating it to create a larger one, ensuring that the training data in the later parts was identical to the beginning. However, the issue still occurred.

Below are some screenshots showing the training time and the status of the devices. The dataset used for training is OPV2V+.

Have you encountered a similar issue during training, or do you have any suggestions on how to resolve this?

Thanks again for your great work!

b73c251a756ac11893b9e3bbb54bd54 a9be31574d9c64d9be8d54bc4ef71cc 5ddbde7d3de3b45c6fb6ddd82e95dd3
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant