Nvidia plan to enhance AI induced success

AI GPU

Nvidia have announced a new generation of artificial intelligence chips and software for running AI models. It’s called: The Blackwell B200 GPU

Blackwell B200 GPU

The Blackwell B200 is the successor to Nvidia’s Hopper H100 and H200 GPUs.

It represents a massive generational leap in computational power.

AI Performance: The B200 GPU delivers 4 times the AI training performance and 30 times the inference performance compared to its predecessor.

Transistor Count: It packs an impressive 208 billion transistors, more than doubling the transistor count of the existing H100.

Memory: The B200 features 192GB of HBM3e memory with an impressive bandwidth of 8 TB/s.

Architecture: The Blackwell architecture takes over from H100/H200.

*Dual-Die Configuration: The B200 is not a single GPU in the traditional sense. Instead, it consists of two tightly coupled die, functioning as one unified CUDA GPU. These chips are linked via a 10 TB/s NV-HBI connection to ensure coherent operation.

*Dual-die packaging technology is used to pack two integrated circuit chips in one single package module. It doubles functionality levels.

Process Node: The B200 utilizes TSMC’s 4NP process node, a refined version of the 4N process used by Hopper H100 and Ada Lovelace architecture GPUs.

The Blackwell B200 is designed for data centres and AI workloads but will likely be available to expect consumer in the future, although these may differ significantly from the data centre model.

Grace Blackwell GB200 Superchip:

Nvidia’s GB200 Grace Blackwell Superchip, with two B200 graphics processors and one Arm-based central processor

This superchip pairs the Grace CPU architecture with the updated Blackwell GPU.

It’s another addition to Nvidia’s lineup, combining CPU and GPU power for advanced computing tasks.

Nvidia continues to push the boundaries of accelerated computing, and these new GPUs promise remarkable performance improvements for AI and other workloads.

Onwards and upwards for Nvidia and the advancement of AI.

Nvidia unveils its newest GH200 high-end AI superchip

Art impression of AI chip

Nvidia has recently announced its latest high-end chip, the GH200 Grace Hopper Superchip, which is designed for training AI models at giant scale. 

The GH200 is a breakthrough accelerated CPU that combines the NVIDIA Grace™ and Hopper™ architectures using NVIDIA® NVLink®-C2C to deliver a CPU+GPU coherent memory model for AI and HPC applications. The superchip delivers up to 10X higher performance for applications running terabytes of data, enabling scientists and researchers to reach unprecedented solutions for the world’s most complex problems.

The technical bit

The GH200 features 72 cores of Grace CPU outfitted with 480 GB of ECC LPDDR5X memory, as well as the GH100 compute GPU that is paired with 141 GB of HBM3E memory that comes in six 24 GB stacks and uses a 6,144-bit memory interface. 

The GH200 also has a new 900 gigabytes per second (GB/s) coherent interface, which is 7X faster than PCIe Gen5, and supercharges accelerated computing and generative AI with HBM3 and HBM3e GPU memory. The GH200 can run all NVIDIA software stacks and platforms, including NVIDIA AI Enterprise, HPC SDK, and Omniverse™.

Nvidia GH200 superchip for AI
Nvidia unveils its newest GH200 high-end AI Superchip.

The GH200 is available as part of the NVIDIA DGX GH200, a massive memory supercomputer that fully connects 256 GH200 Superchips into a singular GPU. The DGX GH200 offers 144 terabytes (TB) of shared memory with linear scalability for giant AI models. 

The DGX GH200 is a turnkey data centre-class solution that includes integrated software and white-glove services from NVIDIA, from design to deployment, to speed the ROI of AI. 

The DGX GH200 is the only AI supercomputer that offers a massive, shared memory space of 144TB across 256 NVIDIA Grace Hopper Superchips, providing developers with nearly 500X more memory to build giant models.

Full details available on the Nvidia website.