A cost and speed comparison between the Lambda Hyperplane 8 V100 GPU Server and AWS p3 GPU instances. A very similar comparison to the DGX-1.
The Lambda Deep Learning Blog
How To Use mpirun to Launch a LLAMA Inference Job Across Multiple Cloud Instances
March 14, 2023
Voltron Data Case Study: Why ML teams are using Lambda Reserved Cloud Clusters
November 01, 2022
How to fine tune stable diffusion: how we made the text-to-pokemon model at Lambda
September 28, 2022
We reproduce the latest Fast.ai/DIUx's ImageNet result with a single 8 Turing GPUs (Titan RTX) server. It takes 2.36 hours to achieve 93% Top-5 accuracy.
This blog tests how fast does ResNet9 (the fastest way to train a SOTA image classifier on Cifar10) run on Nvidia's Turing GPUs, including 2080 Ti and Titan RTX. We also include 1080 Ti as the baseline for comparison.
This blog gives a brief introduction on the history of object detection, explains the idea behind Single-Shot Detection (SSD), and discusses a number of implementation details that will make-or-break the performance. Tensorflow implementation is also provided.