Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

GPU Memory-Usage goes higher during Trainning. #51

Open
yuanryann opened this issue Jul 6, 2024 · 1 comment
Open

GPU Memory-Usage goes higher during Trainning. #51

yuanryann opened this issue Jul 6, 2024 · 1 comment

Comments

@yuanryann
Copy link

Hi Dr.Zhou, Firstly, Thank you very much for your excellent work!!!.
I run trainning on my Nvidia-A4000 GPU,GPU Memory-Usage goes higher every epoch.
Some hparams are changed:
the train_batch_size and the val_batch_size set as 64. parallel :True,num_workers:6,pin_memory: false
Epoch 0, the GPU Memory-Usage takes up 11000Mib, but Epoch 30, it takes up 15577Mib.
Could Anyone help me deal with this issue.
图片

the hparams.yaml shown as below:
historical_steps: 20
future_steps: 30
num_modes: 6
rotate: true
node_dim: 2
edge_dim: 2
embed_dim: 64
num_heads: 8
dropout: 0.1
num_temporal_layers: 4
num_global_layers: 3
local_radius: 50
parallel: true
lr: 0.0005
weight_decay: 0.0001
T_max: 64
root: /home/com0179/AI/Prediction/HiVT/datasets
train_batch_size: 64
val_batch_size: 64
shuffle: true
num_workers: 6
pin_memory: false
persistent_workers: true
gpus: 1
max_epochs: 64
monitor: val_minFDE
save_top_k: 5

@Joseph-Lee-V
Copy link

Hello, I encountered the same issue :)
Have you resolved it?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants