Chuan Li, PhD reviews GPT-3, the new NLP model from OpenAI. The technical overview covers how GPT-3 was trained, GPT-2 vs. GPT-3, and GPT-3 performance.
The Lambda Deep Learning Blog
Featured Posts
Categories
- gpu-cloud (23)
- tutorials (23)
- benchmarks (21)
- announcements (14)
- lambda cloud (13)
- hardware (11)
- NVIDIA H100 (10)
- tensorflow (9)
- gpus (8)
- NVIDIA A100 (7)
- deep learning (6)
- hyperplane (6)
- training (6)
- LLMs (5)
- company (5)
- gpu clusters (5)
- CNNs (4)
- generative networks (4)
- news (4)
- presentation (4)
- rtx a6000 (4)
Recent Posts
OpenAI recently published a blog post on their GPT-2 language model. This tutorial shows you how to run the text generator code yourself.
Published 02/16/2019 by Stephen Balaban
BERT is Google's SOTA pre-training language representations. This blog is about running BERT with multiple GPUs. Specifically, we will use the Horovod framework to parrallelize the tasks. We ill list all the changes to the original BERT implementation and highlight a few places that will make or break the performance.
Published 02/06/2019 by Chuan Li
...