This blog will walk you through the steps of setting up a Horovod [https://github.com/horovod/horovod] + Keras [https://keras.io/] environment for multi-GPU training. Prerequisite * Hardware: A machine with at least two GPUs * Basic Software: Ubuntu (18.04 or 16.04), Nvidia Driver (418.43), CUDA (10.0)
The Lambda Deep Learning Blog
Voltron Data Case Study: Why ML teams are using Lambda Reserved Cloud Clusters
November 01, 2022
How to fine tune stable diffusion: how we made the text-to-pokemon model at Lambda
September 28, 2022
This presentation is a high-level overview of the different types of training regimes that you'll encounter as you move from single GPU to multi GPU to multi node distributed training. It briefly describes where the computation happens, how the gradients are communicated, and how the models are updated and communicated.
BERT is Google's SOTA pre-training language representations. This blog is about running BERT with multiple GPUs. Specifically, we will use the Horovod framework to parrallelize the tasks. We ill list all the changes to the original BERT implementation and highlight a few places that will make or break the performance.