-
Notifications
You must be signed in to change notification settings - Fork 161
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
RuntimeError when resuming training #66
Comments
Hello, I have met the same problem and do you solve it? Thank you very much! |
Hi, I didn't solve it sorry |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Hi, I was training on my own dataset, but when I run the code train.py with --resume option I got this error:
Traceback (most recent call last):
File "tools/train.py", line 239, in
trainer.train()
File "tools/train.py", line 147, in train
self.optimizer.step()
File "/home/cpiedrahita/anaconda3/envs/segmentron/lib/python3.6/site-packages/torch/optim/lr_scheduler.py", line 66, in wrapper
return wrapped(*args, **kwargs)
File "/home/cpiedrahita/anaconda3/envs/segmentron/lib/python3.6/site-packages/torch/optim/sgd.py", line 106, in step
p.data.add_(-group['lr'], d_p)
RuntimeError: value cannot be converted to type float without overflow: (6.33039e-07,-2.05687e-07)
My environment: python 3.6, pytorch 1.4, cuda 10.1
Thanks!
The text was updated successfully, but these errors were encountered: