Lambda secured a $44 million Series B to accelerate the growth of our AI cloud. Funds will be used to deploy new H100 GPU capacity with high-speed network interconnects and develop features that will make Lambda the best cloud in the world for training AI.
The Lambda Deep Learning Blog
How To Use mpirun to Launch a LLAMA Inference Job Across Multiple Cloud Instances
March 14, 2023
Voltron Data Case Study: Why ML teams are using Lambda Reserved Cloud Clusters
November 01, 2022
How to fine tune stable diffusion: how we made the text-to-pokemon model at Lambda
September 28, 2022
In early April, NVIDIA H100 Tensor Core GPUs, the fastest GPU type on the market, will be added to Lambda Cloud. NVIDIA H100 80GB PCIe Gen5 instances will go live first, with SXM to follow very shortly after.
How to use mpirun to launch a LLAMA inference job across multiple cloud instances if you do not have a multi-GPU workstation or server. Despite being more memory efficient than previous language foundation models, LLAMA still requires multiple GPUs to run inference with.
Lambda's GPU cloud has a new team feature that allows you to invite your team to join your account for easy collaboration and more.
Native support for FP8 data types is here with the release of the NVIDIA H100 Tensor Core GPU. These new FP8 types can speed up training and inference.
Lambda and Hugging Face are collaborating on a 2-week sprint to fine-tune OpenAI's Whisper model in as many languages as possible.
This blog walks through how to fine tune stable diffusion to create a text-to-naruto character model, emphasizing the importance of "prompt engineering". Try it out yourself or use it to learn how to train your own Stable Diffusion variants.
In this blog, we will outline the benefits of our new Reserved Cloud Cluster and an example of how Voltron Data is using it to work with large datasets.
In this blog post, we benchmark RTX 4090 to assess its deep learning training performance and compare its performance against RTX 3090, the flagship consumer GPU of the previous Ampere generation.
In this blog post, we go over the most recent updates we made to Lambda on-demand GPU cloud in September, 2022.
Lambda presents an inference benchmark of Stable Diffusion model with different GPUs and CPUs.
Discuss the performance and scalability of H100 GPUs and the whys for upgrading your ML infrastructure with this upcoming big release from NVIDIA.
It can require a lot of work and prompt engineering to get results in a particular style using Stable Diffusion. Fine tuning the image generation model itself presents an efficient solution. This blog walks through how to fine tune Stable Diffusion on a Pokemon dataset to create a text to Pokemon image model.