Lambda has launched a new Hyperplane server combining the fastest GPU on the market, NVIDIA H100, with the world’s best data center CPU, AMD EPYC 9004.
The Lambda Deep Learning Blog
Featured Posts
Categories
- gpu-cloud (23)
- tutorials (23)
- benchmarks (21)
- announcements (14)
- lambda cloud (13)
- hardware (11)
- NVIDIA H100 (10)
- tensorflow (9)
- gpus (8)
- NVIDIA A100 (7)
- deep learning (6)
- hyperplane (6)
- training (6)
- LLMs (5)
- company (5)
- gpu clusters (5)
- CNNs (4)
- generative networks (4)
- news (4)
- presentation (4)
- rtx a6000 (4)
Recent Posts
This post discusses the Total Cost of Ownership (TCO) for a variety of Lambda A100 servers and clusters. We calculate the TCO for individual Hyperplane-A100 servers, compare the cost with renting a AWS p4d.24xlarge instance, and walk through the cost of building and operating A100 clusters.
Published 09/22/2021 by Chuan Li
Introducing the Lambda Echelon, a GPU cluster designed for AI. It comes with the compute, storage, network, power, and support you need to tackle large scale deep learning tasks. Echelon offers a turn-key solution to faster training, faster hyperparameter search, and faster inference.
Published 10/06/2020 by Stephen Balaban
This post uses our Total Cost of Ownership (TCO) calculator to examine the cost of a variety of Lambda Hyperplane-16 clusters. We have the option to include 100 Gb/s EDR InfiniBand networking, storage servers, and complete rack-stack-label-cable service.
Published 04/07/2020 by Stephen Balaban
Scaling out deep learning infrastructure becomes easier with 16 NVIDIA Tesla V100 GPUs and preinstalled frameworks like TensorFlow, Keras, and PyTorch.
Published 12/19/2019 by Chuan Li
A cost and speed comparison between the Lambda Hyperplane 8 V100 GPU Server and AWS p3 GPU instances. A very similar comparison to the DGX-1.
Published 02/11/2019 by Chuan Li