NVIDIABl Chip
NVIDIABl Chip
Computing
New Blackwell GPU, NVLink and Resilience Technologies Enable Trillion-Parameter-Scale AI Models
New Tensor Cores and TensorRT- LLM Compiler Reduce LLM Inference Operating Cost and Energy by up to 25x
New Accelerators Enable Breakthroughs in Data Processing, Engineering Simulation, Electronic Design Automation,
Computer-Aided Drug Design and Quantum Computing
Widespread Adoption by Every Major Cloud Provider, Server Maker and Leading AI Company
GTC—Powering a new era of computing, NVIDIA today announced that the NVIDIA Blackwell platform has arrived —
enabling organizations everywhere to build and run real-time generative AI on trillion-parameter large language models at up
to 25x less cost and energy consumption than its predecessor.
The Blackwell GPU architecture features six transformative technologies for accelerated computing, which will help unlock
breakthroughs in data processing, engineering simulation, electronic design automation, computer-aided drug design,
quantum computing and generative AI — all emerging industry opportunities for NVIDIA.
“For three decades we’ve pursued accelerated computing, with the goal of enabling transformative breakthroughs like deep
learning and AI,” said Jensen Huang, founder and CEO of NVIDIA. “Generative AI is the defining technology of our time.
Blackwell is the engine to power this new industrial revolution. Working with the most dynamic companies in the world, we
will realize the promise of AI for every industry.”
Among the many organizations expected to adopt Blackwell are Amazon Web Services, Dell Technologies, Google, Meta,
Microsoft, OpenAI, Oracle, Tesla and xAI.
Sundar Pichai, CEO of Alphabet and Google: “Scaling services like Search and Gmail to billions of users has taught us a
lot about managing compute infrastructure. As we enter the AI platform shift, we continue to invest deeply in infrastructure for
our own products and services, and for our Cloud customers. We are fortunate to have a longstanding partnership with
NVIDIA, and look forward to bringing the breakthrough capabilities of the Blackwell GPU to our Cloud customers and teams
across Google, including Google DeepMind, to accelerate future discoveries.”
Andy Jassy, president and CEO of Amazon: “Our deep collaboration with NVIDIA goes back more than 13 years, when
we launched the world’s first GPU cloud instance on AWS. Today we offer the widest range of GPU solutions available
anywhere in the cloud, supporting the world’s most technologically advanced accelerated workloads. It's why the new
NVIDIA Blackwell GPU will run so well on AWS and the reason that NVIDIA chose AWS to co-develop Project Ceiba,
combining NVIDIA’s next-generation Grace Blackwell Superchips with the AWS Nitro System's advanced virtualization and
ultra-fast Elastic Fabric Adapter networking, for NVIDIA's own AI research and development. Through this joint effort between
AWS and NVIDIA engineers, we're continuing to innovate together to make AWS the best place for anyone to run NVIDIA
GPUs in the cloud.”
Michael Dell, founder and CEO of Dell Technologies: “Generative AI is critical to creating smarter, more reliable and
efficient systems. Dell Technologies and NVIDIA are working together to shape the future of technology. With the launch of
Blackwell, we will continue to deliver the next-generation of accelerated products and services to our customers, providing
them with the tools they need to drive innovation across industries.”
Demis Hassabis, cofounder and CEO of Google DeepMind: “The transformative potential of AI is incredible, and it will
help us solve some of the world’s most important scientific problems. Blackwell’s breakthrough technological capabilities will
provide the critical compute needed to help the world’s brightest minds chart new scientific discoveries.”
Mark Zuckerberg, founder and CEO of Meta: “AI already powers everything from our large language models to our content
recommendations, ads, and safety systems, and it's only going to get more important in the future. We're looking forward to
using NVIDIA's Blackwell to help train our open-source Llama models and build the next generation of Meta AI and
consumer products.”
Satya Nadella, executive chairman and CEO of Microsoft: “We are committed to offering our customers the most
advanced infrastructure to power their AI workloads. By bringing the GB200 Grace Blackwell processor to our datacenters
globally, we are building on our long-standing history of optimizing NVIDIA GPUs for our cloud, as we make the promise of
AI real for organizations everywhere.”
Sam Altman, CEO of OpenAI: “Blackwell offers massive performance leaps, and will accelerate our ability to deliver
leading-edge models. We’re excited to continue working with NVIDIA to enhance AI compute.”
Larry Ellison, chairman and CTO of Oracle: "Oracle’s close collaboration with NVIDIA will enable qualitative and
quantitative breakthroughs in AI, machine learning and data analytics. In order for customers to uncover more actionable
insights, an even more powerful engine like Blackwell is needed, which is purpose-built for accelerated computing and
generative AI.”
Elon Musk, CEO of Tesla and xAI: “There is currently nothing better than NVIDIA hardware for AI.”
Named in honor of David Harold Blackwell — a mathematician who specialized in game theory and statistics, and the first
Black scholar inducted into the National Academy of Sciences — the new architecture succeeds the NVIDIA Hopper™
architecture, launched two years ago.
World’s Most Powerful Chip — Packed with 208 billion transistors, Blackwell-architecture GPUs are manufactured
using a custom-built 4NP TSMC process with two-reticle limit GPU dies connected by 10 TB/second chip-to-chip link
into a single, unified GPU.
Second-Generation Transformer Engine — Fueled by new micro-tensor scaling support and NVIDIA’s advanced
dynamic range management algorithms integrated into NVIDIA TensorRT™-LLM and NeMo Megatron frameworks,
Blackwell will support double the compute and model sizes with new 4-bit floating point AI inference capabilities.
Fifth-Generation NVLink — To accelerate performance for multitrillion-parameter and mixture-of-experts AI models,
the latest iteration of NVIDIA NVLink® delivers groundbreaking 1.8TB/s bidirectional throughput per GPU, ensuring
seamless high-speed communication among up to 576 GPUs for the most complex LLMs.
RAS Engine — Blackwell-powered GPUs include a dedicated engine for reliability, availability and serviceability.
Additionally, the Blackwell architecture adds capabilities at the chip level to utilize AI-based preventative maintenance
to run diagnostics and forecast reliability issues. This maximizes system uptime and improves resiliency for massive-
scale AI deployments to run uninterrupted for weeks or even months at a time and to reduce operating costs.
Secure AI — Advanced confidential computing capabilities protect AI models and customer data without compromising
performance, with support for new native interface encryption protocols, which are critical for privacy-sensitive
industries like healthcare and financial services.
Decompression Engine — A dedicated decompression engine supports the latest formats, accelerating database
queries to deliver the highest performance in data analytics and data science. In the coming years, data processing, on
which companies spend tens of billions of dollars annually, will be increasingly GPU-accelerated.
A Massive Superchip
The NVIDIA GB200 Grace Blackwell Superchip connects two NVIDIA B200 Tensor Core GPUs to the NVIDIA Grace CPU
over a 900GB/s ultra-low-power NVLink chip-to-chip interconnect.
For the highest AI performance, GB200-powered systems can be connected with the NVIDIA Quantum-X800 InfiniBand and
Spectrum™-X800 Ethernet platforms, also announced today, which deliver advanced networking at speeds up to 800Gb/s.
The GB200 is a key component of the NVIDIA GB200 NVL72, a multi-node, liquid-cooled, rack-scale system for the most
compute-intensive workloads. It combines 36 Grace Blackwell Superchips, which include 72 Blackwell GPUs and 36 Grace
CPUs interconnected by fifth-generation NVLink. Additionally, GB200 NVL72 includes NVIDIA BlueField®-3 data
processing units to enable cloud network acceleration, composable storage, zero-trust security and GPU compute elasticity
in hyperscale AI clouds. The GB200 NVL72 provides up to a 30x performance increase compared to the same number of
NVIDIA H100 Tensor Core GPUs for LLM inference workloads, and reduces cost and energy consumption by up to 25x.
The platform acts as a single GPU with 1.4 exaflops of AI performance and 30TB of fast memory, and is a building block for
the newest DGX SuperPOD.
NVIDIA offers the HGX B200, a server board that links eight B200 GPUs through NVLink to support x86-based generative AI
platforms. HGX B200 supports networking speeds up to 400Gb/s through the NVIDIA Quantum-2 InfiniBand and Spectrum-X
Ethernet networking platforms.
AWS, Google Cloud, Microsoft Azure and Oracle Cloud Infrastructure will be among the first cloud service providers to offer
Blackwell-powered instances, as will NVIDIA Cloud Partner program companies Applied Digital, CoreWeave, Crusoe, IBM
Cloud and Lambda. Sovereign AI clouds will also provide Blackwell-based cloud services and infrastructure, including
Indosat Ooredoo Hutchinson, Nebius, Nexgen Cloud, Oracle EU Sovereign Cloud, the Oracle US, UK, and Australian
Government Clouds, Scaleway, Singtel, Northern Data Group's Taiga Cloud, Yotta Data Services’ Shakti Cloud and YTL
Power International.
GB200 will also be available on NVIDIA DGX™ Cloud, an AI platform co-engineered with leading cloud service providers
that gives enterprise developers dedicated access to the infrastructure and software needed to build and deploy advanced
generative AI models. AWS, Google Cloud and Oracle Cloud Infrastructure plan to host new NVIDIA Grace Blackwell-based
instances later this year.
Cisco, Dell, Hewlett Packard Enterprise, Lenovo and Supermicro are expected to deliver a wide range of servers based on
Blackwell products, as are Aivres, ASRock Rack, ASUS, Eviden, Foxconn, GIGABYTE, Inventec, Pegatron, QCT,
Wistron, Wiwynn and ZT Systems.
Additionally, a growing network of software makers, including Ansys, Cadence and Synopsys — global leaders in
engineering simulation — will use Blackwell-based processors to accelerate their software for designing and simulating
electrical, mechanical and manufacturing systems and parts. Their customers can use generative AI and accelerated
computing to bring products to market faster, at lower cost and with higher energy efficiency.
To learn more about the NVIDIA Blackwell platform, watch the GTC keynote and register to attend sessions from NVIDIA and
industry leaders at GTC, which runs through March 21.
About NVIDIA
Since its founding in 1993, NVIDIA (NASDAQ: NVDA) has been a pioneer in accelerated computing. The company’s
invention of the GPU in 1999 sparked the growth of the PC gaming market, redefined computer graphics, ignited the era of
modern AI and is fueling industrial digitalization across markets. NVIDIA is now a full-stack computing infrastructure
company with data-center-scale offerings that are reshaping industry. More information at https://nvidianews.nvidia.com/.
Certain statements in this press release including, but not limited to, statements as to: the benefits, impact, performance,
features, and availability of NVIDIA’s products and technologies, including NVIDIA Blackwell platform, Blackwell GPU
architecture, Resilience Technologies, Custom Tensor Core technology, NVIDIA TensorRT-LLM, NeMo Megatron
framework, NVLink, NVIDIA GB200 Grace Blackwell Superchip, B200 Tensor Core GPUs, NVIDIA Grace CPU, NVIDIA
H100 Tensor Core GPU, NVIDIA Quantum-X800 InfiniBand and Spectrum-X800 Ethernet platforms, NVIDIA GB200 NVL72,
NVIDIA BlueField-3 data processing units, DGX SuperPOD, HGX B200, Quantum-2 InfiniBand and Spectrum-X Ethernet
platforms, BlueField-3 DPUs, NVIDIA DGX Cloud, NVIDIA AI Enterprise, and NVIDIA NIM inference microservices; our goal
of enabling transformative breakthroughs like deep learning and AI; Blackwell GPUs being the engine to power a new
industrial revolution; our ability to realize the promise of AI for every industry as we working with the most dynamic
companies in the world; our collaborations and partnerships with third parties and the benefits and impacts thereof; third
parties who will offer or use our products, services and infrastructures and who will deliver servers based on our products;
and the ability of the customers of global leaders in engineering simulation to use generative AI and accelerated computing
to bring products to market faster, at lower cost and with higher energy efficiency are forward-looking statements that are
subject to risks and uncertainties that could cause results to be materially different than expectations. Important factors that
could cause actual results to differ materially include: global economic conditions; our reliance on third parties to
manufacture, assemble, package and test our products; the impact of technological development and competition;
development of new products and technologies or enhancements to our existing product and technologies; market
acceptance of our products or our partners' products; design, manufacturing or software defects; changes in consumer
preferences or demands; changes in industry standards and interfaces; unexpected loss of performance of our products or
technologies when integrated into systems; as well as other factors detailed from time to time in the most recent reports
NVIDIA files with the Securities and Exchange Commission, or SEC, including, but not limited to, its annual report on Form
10-K and quarterly reports on Form 10-Q. Copies of reports filed with the SEC are posted on the company's website and are
available from NVIDIA without charge. These forward-looking statements are not guarantees of future performance and
speak only as of the date hereof, and, except as required by law, NVIDIA disclaims any obligation to update these forward-
looking statements to reflect future events or circumstances.
© 2024 NVIDIA Corporation. All rights reserved. NVIDIA, the NVIDIA logo, BlueField, DGX, NVIDIA HGX, NVIDIA Hopper,
NVIDIA NeMo, NVIDIA NIM, NVIDIA Spectrum, NVLink, and TensorRT are trademarks and/or registered trademarks of
NVIDIA Corporation in the U.S. and other countries. Other company and product names may be trademarks of the respective
companies with which they are associated. Features, pricing, availability and specifications are subject to change without
notice.
Kristin Uchiyama
Enterprise and Edge Computing
+1-408-486-2248
kuchiyama@nvidia.com