Upcoming Talks

Ista white

Decentralized Deep Learning: Training and Running Large Models over the Internet

Date
Tuesday, March 21, 2023 13:00 - 15:00
Speaker
Max Ryabinin
Location
Heinzel Seminar Room / Office Bldg West (I21.EG.101)
Series
Seminar/Talk
Tags
CS Talk Series
Host
Dan Alistarh
Contact

Over the recent years, the scale of deep learning has increased dramatically: pretraining models like GPT-3 can cost millions of dollars, and even their inference requires significant resources. In this talk, I will present an alternative approach: instead of using expensive clusters, we can leverage the resources of volunteers or several organizations. I will cover several papers addressing the challenges of such a setup published at NeurIPS'20, '21, and ICML'22, as well as Hivemind our open-source library for decentralized DL.

I will also highlight SWARM Parallelism and Petals our latest works about decentralized pretraining and inference of large language models. SWARM is a system for training large models over slow networks of heterogeneous unreliable devices: to achieve this goal, it relies on randomized pipelines and dynamic rebalancing between pipeline stages. In turn, Petals leverages the same techniques and allows everyone to run, finetune or inspect the internals of LLMs like BLOOM or OPT-175B without access to many GPUs or the need for offloading even from Colab notebooks.
Qr image
Download ICS Download invitation
Back to eventlist