Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

dist_train keep waiting #10

Open
FX-STAR opened this issue Jan 26, 2021 · 4 comments
Open

dist_train keep waiting #10

FX-STAR opened this issue Jan 26, 2021 · 4 comments

Comments

@FX-STAR
Copy link

FX-STAR commented Jan 26, 2021

My env:
cuda10.2
torch==1.6.0
mmdetection==2.8.0
mmcv==1.2.4
After some iters the GPU-Util 100% but the process is always waiting
Could you provide your env or any advice?

@hyz-xmaster
Copy link
Owner

hyz-xmaster commented Jan 27, 2021

Hi, I didn't run into this problem so can't provide effective solutions. The code is tested with: cuda 10.1, pytorch 1.6.0, mmdet 2.5.0, and mmcv=1.1.5. You may have a look at this page for more information about training.

@oym050922021
Copy link

@whoNamedCody ,hi,has the problem been resolved? I also faced this problem.

@FX-STAR
Copy link
Author

FX-STAR commented Mar 9, 2021

@whoNamedCody ,hi,has the problem been resolved? I also faced this problem.

no, i think the problem maybe in 'GiouLoss', but not debug yet

@oym050922021
Copy link

oym050922021 commented Mar 11, 2021 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants