Submitted by genuinelySurprised t3_zky7ly in MachineLearning
Given that well-funded groups like Google, Meta and OpenAI may eventually develop an insurmountable lead for services like image classification and NLP that seem to require huge numbers of parameters, I'd be surprised if there wasn't an effort underway to make a BOINC-powered distributed system that millions of us mere peons could contribute to collaboratively. But aside from the now-defunct MLC@Home project, I haven't found anything yet. Am I missing something?
justheuristic t1_j02g9m0 wrote
https://github.com/bigscience-workshop/petals - fine-tuning BLOOM-176B Folding@home style
https://github.com/learning-at-home/hivemind - a library for decentralized training with volunteers
https://github.com/epfml/disco - a library for collaborative training in JS (in a browser!)
https://github.com/chavinlo/distributed-diffusion - a project that tries to train diffusion this way
https://bittensor.com/ - a comminity that makes decentralized training into a cryptocurrency
There are also projects like Together that build networks from university computers for decentralized training.