NVIDIA Investor Presentation slide image

NVIDIA Investor Presentation

Training Compute (petaFLOPS) 1010 109 Modern Al is a Data Center Scale Computing Workload Data centers are becoming Al factories: data as input, intelligence as output Al Training Computational Requirements All Al Models Excluding Transformers: 8x/2yrs Transformer Al Models: 275x/2yrs Megatron-Turing NLG 530B GPT-3 108 Microsoft T-NLG GPT-2 107 Megatron Wav2Vec 2.0 XLNet❤ MoCo ResNet50 106 Xception Inception V3 BERT Large 105 Resnet Seq2Seq GPT-1 Transformer 104 ResNext VGG19 ELMO DenseNet201 103 AlexNet 102 2012 2013 2014 2015 2016 2017 2018 2019 2020 2021 2022 8998 Fueling Giant-Scale Al Infrastructure NVIDIA compute & networking GPU | DPU | CPU 8998 888. 8818. NVIDIA 11111 8818. 11111 Large Language Models, based on the Transformer architecture, are one of today's most important advanced Al technologies, involving up to trillions of parameters that learn from text. Developing them is an expensive, time-consuming process that demands deep technical expertise, distributed data center-scale infrastructure, and a full-stack accelerated computing approach. NVIDIA.
View entire presentation