Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CUDA out of memory #42

Open
KotlinWang opened this issue Oct 6, 2023 · 9 comments
Open

CUDA out of memory #42

KotlinWang opened this issue Oct 6, 2023 · 9 comments

Comments

@KotlinWang
Copy link

Hello, very good work, I used a single 3090 to train MonoDETR and the "CUDA out of memory" prompt appeared. All my configurations use the default monodetr.yaml settings, and my environment configuration is also in accordance with the requirements of README.md, but what is the reason for such a problem during the training phase? Very much looking forward to your reply, thank you!

@charmeleonz
Copy link

Same problem encountered.

@KotlinWang
Copy link
Author

I can only set the batch size to 14 using a single 3090 graphics card, and the network training is very unstable.

@yjy4231
Copy link

yjy4231 commented Oct 15, 2023

Same problem encountered!

@KotlinWang
Copy link
Author

KotlinWang commented Oct 16, 2023 via email

@yjy4231
Copy link

yjy4231 commented Oct 16, 2023

Car [email protected], 0.70, 0.70:
bbox AP:90.4341, 88.1947, 79.9611
bev AP:39.2351, 30.7552, 26.5470
3d AP:28.5708, 22.4689, 20.4412
aos AP:89.67, 86.37, 77.71
Car [email protected], 0.70, 0.70:
bbox AP:96.1279, 89.7959, 82.5666
bev AP:37.3690, 26.5359, 22.8405
3d AP:26.4230, 19.8301, 16.8303
aos AP:95.24, 87.83, 80.01
Car [email protected], 0.50, 0.50:
bbox AP:90.4341, 88.1947, 79.9611
bev AP:71.6413, 53.7894, 47.6267
3d AP:65.7944, 48.1693, 45.8162
aos AP:89.67, 86.37, 77.71
Car [email protected], 0.50, 0.50:
bbox AP:96.1279, 89.7959, 82.5666
bev AP:71.5228, 52.7067, 46.6121
3d AP:67.7813, 48.3621, 43.4522
aos AP:95.24, 87.83, 80.01

I only get the AP40 result of Mod. level is 19.81.

@KotlinWang
Copy link
Author

Car [email protected], 0.70, 0.70: bbox AP:90.4341, 88.1947, 79.9611 bev AP:39.2351, 30.7552, 26.5470 3d AP:28.5708, 22.4689, 20.4412 aos AP:89.67, 86.37, 77.71 Car [email protected], 0.70, 0.70: bbox AP:96.1279, 89.7959, 82.5666 bev AP:37.3690, 26.5359, 22.8405 3d AP:26.4230, 19.8301, 16.8303 aos AP:95.24, 87.83, 80.01 Car [email protected], 0.50, 0.50: bbox AP:90.4341, 88.1947, 79.9611 bev AP:71.6413, 53.7894, 47.6267 3d AP:65.7944, 48.1693, 45.8162 aos AP:89.67, 86.37, 77.71 Car [email protected], 0.50, 0.50: bbox AP:96.1279, 89.7959, 82.5666 bev AP:71.5228, 52.7067, 46.6121 3d AP:67.7813, 48.3621, 43.4522 aos AP:95.24, 87.83, 80.01

I only get the AP40 result of Mod. level is 19.81.

Hello, may I know your graphics device model?

@yjy4231
Copy link

yjy4231 commented Oct 16, 2023

Car [email protected], 0.70, 0.70: bbox AP:90.4341, 88.1947, 79.9611 bev AP:39.2351, 30.7552, 26.5470 3d AP:28.5708, 22.4689, 20.4412 aos AP:89.67, 86.37, 77.71 Car [email protected], 0.70, 0.70: bbox AP:96.1279, 89.7959, 82.5666 bev AP:37.3690, 26.5359, 22.8405 3d AP:26.4230, 19.8301, 16.8303 aos AP:95.24, 87.83, 80.01 Car [email protected], 0.50, 0.50: bbox AP:90.4341, 88.1947, 79.9611 bev AP:71.6413, 53.7894, 47.6267 3d AP:65.7944, 48.1693, 45.8162 aos AP:89.67, 86.37, 77.71 Car [email protected], 0.50, 0.50: bbox AP:96.1279, 89.7959, 82.5666 bev AP:71.5228, 52.7067, 46.6121 3d AP:67.7813, 48.3621, 43.4522 aos AP:95.24, 87.83, 80.01
I only get the AP40 result of Mod. level is 19.81.

Hello, may I know your graphics device model?

a single 3090 GPU with batch_size=14

@Ivan-Tang-3D
Copy link
Collaborator

The original version is for the 3090, while the stable version is for the A100. With the skill of Group DETR, the cuda memory could reach 40G.

@Ivan-Tang-3D
Copy link
Collaborator

If u want to adapt the model to 3090, u could set the group_detr param in cfg to 1,and comment the lines of 467-473(about conditional) in the https://github.com/ZrrSkywalker/MonoDETR/blob/main/lib/models/monodetr/depthaware_transformer.py, then the model turns to the original version.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants