**hivemind.optim**
==================
.. raw:: html
This module contains decentralized optimizers that wrap regular pytorch optimizers to collaboratively train a shared model. Depending on the exact type, optimizer may average model parameters with peers, exchange gradients, or follow a more complicated distributed training strategy.
.. automodule:: hivemind.optim.experimental.optimizer
.. currentmodule:: hivemind.optim.experimental.optimizer
**hivemind.Optimizer**
----------------------
.. autoclass:: Optimizer
:members: step, zero_grad, load_state_from_peers, param_groups, shutdown
:member-order: bysource
.. currentmodule:: hivemind.optim.grad_scaler
.. autoclass:: GradScaler
:member-order: bysource
**CollaborativeOptimizer**
--------------------------
.. raw:: html
CollaborativeOptimizer is a legacy version of hivemind.Optimizer. **For new projects, please use hivemind.Optimizer.**
Currently, hivemind.Optimizer supports all the features of CollaborativeOptimizer and then some.
CollaborativeOptimizer will still be supported for awhile, but will eventually be deprecated.
.. automodule:: hivemind.optim.collaborative
.. currentmodule:: hivemind.optim
.. autoclass:: CollaborativeOptimizer
:members: step
:member-order: bysource
.. autoclass:: CollaborativeAdaptiveOptimizer
:members:
:member-order: bysource