Generative AI Computing Platform slide image

Generative AI Computing Platform

1010 109 108 Training Compute (petaFLOPs) 107 106 Modern Al is a Data Center Scale Computing Workload Data centers are becoming Al factories: Data as input, intelligence as output Before Transformers = 8x/2yrs Transformers = 215x/2yrs 105 104 103 ⚫ AlexNet MT NLG 530B .. GPT-3⚫ Microsoft T-NLG GPT-2⚫ Megatron-NLG⚫ PaLM • Chinchilla XLNet⚫ Wav2Vec 2.0 MoCo ResNet50 Xception BERT Large Inception V3 ⚫ Seq2Seq Resnet VGG-19 • GPT-1 Transformer • ResNeXt • ELMO DenseNet201 ● • BLOOM 102 2011 2012 2013 2014 2015 2016 2017 2018 2019 2020 2021 2022 2023 Al Training Computational Requirements 9898. 88988 9898. NVIDIA nVIDI nVIDIA Large Language Models, based on the Transformer architecture, are one of today's most important advanced Al technologies, involving up to trillions of parameters that learn from text. 8898. 8898. 8898. 8898 8898. Developing them is an expensive, time-consuming process that demands deep technical expertise, distributed data center-scale infrastructure, and a full-stack accelerated computing approach. NVIDI 100 -ηνικ Grvou Greens NVIDIA Fueling Giant-Scale Al Infrastructure NVIDIA compute & networking GPU | DPU | CPU 1004 12 NVIDIA.
View entire presentation