You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi,
Thanks for your great work. I ran rocl_train.py with your default parameters and find one error output when executing loss.backward(), the error is: _RuntimeError: one of the variables needed for gradient computation has been modified by an inplace operation.
I found the problem is in adv_loss. After we execute outputs = projector(model(inputs)), adv_loss cannot be backwarded due to the inplace operation. Appreciate for your help.
The text was updated successfully, but these errors were encountered:
It runs fine with multiple gpus. Error pops up for single gpu run. Comment out the lines 88 to 100 which wraps the model with SyncBN & DDP. Also the init (line 47 to 52, although I haven't tried with keeping this intact). Accordingly, update the run command by removing the distributed part. It should work fine then.
Hi,
Thanks for your great work. I ran rocl_train.py with your default parameters and find one error output when executing
loss.backward()
, the error is:_RuntimeError: one of the variables needed for gradient computation has been modified by an inplace operation.
I found the problem is in
adv_loss
. After we executeoutputs = projector(model(inputs))
,adv_loss
cannot be backwarded due to the inplace operation. Appreciate for your help.The text was updated successfully, but these errors were encountered: