A cost and speed comparison between the Lambda Hyperplane 8 V100 GPU Server and AWS p3 GPU instances. A very similar comparison to the DGX-1.
The Lambda Deep Learning Blog
Voltron Data Case Study: Why ML teams are using Lambda Reserved Cloud Clusters
November 01, 2022
How to fine tune stable diffusion: how we made the text-to-pokemon model at Lambda
September 28, 2022
Lambda Cloud Storage is now in open beta: a high speed filesystem for our GPU instances
April 18, 2022
Lambda Teams Up With Razer to Launch the World’s Most Powerful Laptop for Deep Learning
April 11, 2022
We reproduce the latest Fast.ai/DIUx's ImageNet result with a single 8 Turing GPUs (Titan RTX) server. It takes 2.36 hours to achieve 93% Top-5 accuracy.
This blog tests how fast does ResNet9 (the fastest way to train a SOTA image classifier on Cifar10) run on Nvidia's Turing GPUs, including 2080 Ti and Titan RTX. We also include 1080 Ti as the baseline for comparison.
This blog gives a brief introduction on the history of object detection, explains the idea behind Single-Shot Detection (SSD), and discusses a number of implementation details that will make-or-break the performance. Tensorflow implementation is also provided.