You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello, I have been trying to reproduce your REFERS code after reading your masterpiece paper. But when I read you codes, I have a question.
You wrap torch.distributed in https://github.com/funnyzhou/REFERS/blob/master/Pre-train/refers/utils/distributed.py, as I understand it, it is a well-wrapped function that can be operated on multiple gpus on one or even multiple machines. But when I tried to run pretrain_refers.py on 2 gpus, the program crushed down after 100 iterations. So my question is how can I run pretrain on multiple gpus?
Hello, I have been trying to reproduce your REFERS code after reading your masterpiece paper. But when I read you codes, I have a question.