justheuristic пре 3 година
родитељ
комит
408e926915
1 измењених фајлова са 1 додато и 1 уклоњено
  1. 1 1
      hivemind/optim/experimental/optimizer.py

+ 1 - 1
hivemind/optim/experimental/optimizer.py

@@ -220,7 +220,7 @@ class Optimizer(torch.optim.Optimizer):
         if self.batch_size_per_step is None and batch_size is None and not self.auxiliary:
             raise ValueError("Please either set batch_size_per_step parameter at init or when calling .step")
         if self.auxiliary and (closure is not None or batch_size is not None or grad_scaler is not None):
-            raise ValueError("Auxiliary peers should not have batch size run closures, or use grad_scaler")
+            raise ValueError("Auxiliary peers should not have batch size, run closures, or use grad_scaler")
         batch_size = batch_size if batch_size is not None else self.batch_size_per_step
 
         loss = None