瀏覽代碼

Add clarification to averaging quickstart (#492)

It is not immediately clear from the documentation that this example cannot run on multiple machines. This PR clarifies this.

(cherry picked from commit ee75b913af17c4f53e5b1b45413ee10629020699)
Adrian Castro 3 年之前
父節點
當前提交
c0e87c0e29
共有 1 個文件被更改,包括 4 次插入3 次删除
  1. 4 3
      docs/user/quickstart.md

+ 4 - 3
docs/user/quickstart.md

@@ -22,8 +22,10 @@ We assume that you are already familiar with the official [CIFAR-10 example](htt
 from the PyTorch website.
 
 We build on top of the official example to spin up distributed training of a two-layer neural network by averaging weights.
-For simplicity, this tutorial will use two non-GPU peers running on the same machine. If you get to the end of this
-tutorial, we'll give you an example of actual distributed training of Transformers ;)
+For simplicity, this tutorial will use two non-GPU peers running on the same machine. If you try to run this example on two
+separate machines with different IPs, this example will not work. To read more about how to perform training on more
+than one machine check out [DHT - Running Across the Internet](https://learning-at-home.readthedocs.io/en/latest/user/dht.html#running-across-the-internet).
+If you get to the end of this tutorial, we'll give you an example of actual distributed training of Transformers ;)
 
 For now, let's run our first training peer:
 ```python
@@ -187,4 +189,3 @@ If you want to learn more about each individual component,
 - Learn the underlying math behind hivemind.Optimizer in [Diskin et al., (2021)](https://arxiv.org/abs/2106.10207), 
   [Li et al. (2020)](https://arxiv.org/abs/2005.00124) and [Ryabinin et al. (2021)](https://arxiv.org/abs/2103.03239).
 - Read about setting up Mixture-of-Experts training in [this guide](https://learning-at-home.readthedocs.io/en/latest/user/moe.html),
-