Hivemind is a PyTorch library to train large neural networks across the Internet. Imagine training one huge Transformer model on thousands of computers from different universities, companies, and volunteers.
To learn more about the idea behind this library and its components, see https://learning-at-home.github.io or read the NeurIPS 2020 paper
Hivemind is currently at the active development stage, and we welcome all contributions from bug fixes and documentation improvements to entirely new features. If you want to contribute to hivemind, take a look at the issues or join our chat room. The Developer's guide page contains best practices, as well as description of tests and performance benchmarks.
You can read the paper that inspired hivemind here:
Towards Crowdsourced Training of Large Neural Networks using Decentralized Mixture-of-Experts (Max Ryabinin and Anton Gusev, NeurIPS 2020).
@misc{ryabinin2020crowdsourced,
title={Towards Crowdsourced Training of Large Neural Networks using Decentralized Mixture-of-Experts},
author={Max Ryabinin and Anton Gusev},
year={2020},
eprint={2002.04013},
archivePrefix={arXiv},
primaryClass={cs.DC}
}
The initial implementation of hivemind used to conduct experiments for the paper is available here: mryab/learning-at-home.
In the docs, we list several related projects and acknowledgements.