You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on May 4, 2020. It is now read-only.
will be interesting to have a look if we observe similar speed, and code is probably useful too.
note that their benchmark is only raw communication all-reduce, no learning. this is relevant if one is communication bound. so we might likely see this scenario when training linear models soon
The text was updated successfully, but these errors were encountered:
we found this benchmark here:
https://github.com/diux-dev/cluster/tree/master/pytorch_distributed_benchmark
will be interesting to have a look if we observe similar speed, and code is probably useful too.
note that their benchmark is only raw communication all-reduce, no learning. this is relevant if one is communication bound. so we might likely see this scenario when training linear models soon
The text was updated successfully, but these errors were encountered: