Benchmarks on NVIDIA’s Transformer Engine, which boosts FP8 performance by an impressive 60% on GPT3-style model testing on NVIDIA H100 Tensor Core GPUs.
The Lambda Deep Learning Blog
Featured Posts
Categories
- gpu-cloud (23)
- tutorials (23)
- benchmarks (21)
- announcements (14)
- lambda cloud (13)
- hardware (11)
- NVIDIA H100 (10)
- tensorflow (9)
- gpus (8)
- NVIDIA A100 (7)
- deep learning (6)
- hyperplane (6)
- training (6)
- LLMs (5)
- company (5)
- gpu clusters (5)
- CNNs (4)
- generative networks (4)
- news (4)
- presentation (4)
- rtx a6000 (4)
Recent Posts
Lambda will be one of the first cloud providers in the world to offer customers access to NVIDIA H200 Tensor Core GPUs through Lambda Cloud Clusters.
Published 11/13/2023 by Maxx Garrison
Lambda Cloud Clusters are now available with the NVIDIA GH200 Grace Hopper Superchip. A single GH200 has 576GB of coherent memory for unmatched efficiency.
Published 11/13/2023 by Maxx Garrison
GPU benchmarks on Lambda’s offering of the NVIDIA H100 SXM5 vs the NVIDIA A100 SXM4 using DeepChat’s 3-step training example.
Published 10/12/2023 by Chuan Li
Persistent storage for Lambda Cloud is expanding. Filesystems are now available for all regions except Utah, which is coming very soon.
Published 09/20/2023 by Kathy Bui
This analysis explores whether a Large Language Model, such as OpenAI's GPT, can assist in helping human readers digest scientific reviews.
Published 09/14/2023 by Xi Tian
Lambda has launched a new Hyperplane server combining the fastest GPU on the market, NVIDIA H100, with the world’s best data center CPU, AMD EPYC 9004.
Published 09/07/2023 by Maxx Garrison
How to use FlashAttention-2 on Lambda Cloud, including H100 vs A100 benchmark results for training GPT-3-style models using the new model.
Published 08/24/2023 by Chuan Li
On-demand HGX H100 systems with 8x NVIDIA H100 SXM instances are now available on Lambda Cloud for only $2.59/hr/GPU.
Published 08/02/2023 by Kathy Bui
Unlock the potential of open-source LLMs by hosting your very own langchain+Falcon+Chroma application. Upload a PDF and engage in Q&A about its contents.
Published 07/24/2023 by Xi Tian
This blog post provides instructions on how to fine tune LLaMA 2 models on Lambda Cloud using a $0.60/hr A10 GPU.
Published 07/20/2023 by Corey Lowman
How to build the GPU infrastructure needed to pretrain LLM and Generative AI models from scratch (e.g. GPT-4, LaMDA, LLaMA, BLOOM).
Published 07/13/2023 by David Hall
Learn how to fine-tune Falcon LLM 7B/40B on a single GPU with LoRA and quantization, enabling data parallelism for linear scaling across multiple GPUs.
Published 06/29/2023 by Xi Tian