Benchmarks on NVIDIA’s Transformer Engine, which boosts FP8 performance by an impressive 60% on GPT3-style model testing on NVIDIA H100 Tensor Core GPUs.
The Lambda Deep Learning Blog
Featured Posts
Categories
- gpu-cloud (23)
- tutorials (23)
- benchmarks (21)
- announcements (14)
- lambda cloud (13)
- hardware (11)
- NVIDIA H100 (10)
- tensorflow (9)
- gpus (8)
- NVIDIA A100 (7)
- deep learning (6)
- hyperplane (6)
- training (6)
- LLMs (5)
- company (5)
- gpu clusters (5)
- CNNs (4)
- generative networks (4)
- news (4)
- presentation (4)
- rtx a6000 (4)
Recent Posts
GPU benchmarks on Lambda’s offering of the NVIDIA H100 SXM5 vs the NVIDIA A100 SXM4 using DeepChat’s 3-step training example.
Published 10/12/2023 by Chuan Li
How to use FlashAttention-2 on Lambda Cloud, including H100 vs A100 benchmark results for training GPT-3-style models using the new model.
Published 08/24/2023 by Chuan Li
In this blog, we will outline the benefits of our new Reserved Cloud Cluster and an example of how Voltron Data is using it to work with large datasets.
Published 11/01/2022 by Lauren Watkins
This post discusses the Total Cost of Ownership (TCO) for a variety of Lambda A100 servers and clusters. We calculate the TCO for individual Hyperplane-A100 servers, compare the cost with renting a AWS p4d.24xlarge instance, and walk through the cost of building and operating A100 clusters.
Published 09/22/2021 by Chuan Li
Introducing the Lambda Echelon, a GPU cluster designed for AI. It comes with the compute, storage, network, power, and support you need to tackle large scale deep learning tasks. Echelon offers a turn-key solution to faster training, faster hyperparameter search, and faster inference.
Published 10/06/2020 by Stephen Balaban
Benchmarks for ResNet-152, Inception v3, Inception v4, VGG-16, AlexNet, SSD300, and ResNet-50 using the NVIDIA A100 GPU and DGX A100 server.
Published 05/22/2020 by Stephen Balaban