You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I note that in the abstract of the paper it mentions it supports distributed-memory execution on cloud/supercomputer systems and is available as open source, however not a single actual example exists in the repo. Since the cost of hessian is significant, is it pratical to use DDP to reduce GPU memory load? Or any examples? Thank you.
The text was updated successfully, but these errors were encountered:
I note that in the abstract of the paper it mentions it supports distributed-memory execution on cloud/supercomputer systems and is available as open source, however not a single actual example exists in the repo. Since the cost of hessian is significant, is it pratical to use DDP to reduce GPU memory load? Or any examples? Thank you.
The text was updated successfully, but these errors were encountered: