Browse Source

Update hivemind/optim/averaged.py

Co-authored-by: Max Ryabinin <mryabinin0@gmail.com>
Anton Sinitsin 4 years ago
parent
commit
26ac1c7299
1 changed files with 1 additions and 1 deletions
  1. 1 1
      hivemind/optim/averaged.py

+ 1 - 1
hivemind/optim/averaged.py

@@ -32,7 +32,7 @@ class DecentralizedOptimizer(DecentralizedOptimizerBase):
     :param average_gradients: whether to average gradients
     :param max_allowed_epoch_difference: if max_epoch has difference with local_epoch more than that, we download state
       from other peer.
-    :param total_steps_in_epoch: how many total steps must be to increase local_epoch by one
+    :param total_steps_in_epoch: the number of optimizer steps for a single training epoch
     :param average_opt_statistics: if specified, average optimizer states with corresponding names in state_dict
     :param scheduler_cls: a function which takes an optimizer and returns a learning rate scheduler
     :param averaging_steps_period: performs averaging after this many optimizer steps