Generative AI Computing Platform slide image

Generative AI Computing Platform

Full-Stack & Data Center Scale Acceleration Drive significant cost savings and workload scaling - Classical Computing — 960 CPU-only servers Application CPU server racks LLM Workload: Bert-Large Training and Inference | CPU Server: Dual-EYPC 7763 | GPU Server: Dual-EPYC 7763 + 8X H100 PCle GPUs Accelerated Computing - 2 GPU servers Application Re-Engineered for Acceleration CUDA-X Acceleration Libraries Magnum 10 25X lower cost 84X better energy-efficiency 13 NVIDIA.
View entire presentation