The curtains have closed on another year of the GPU Technology Conference (GTC24) in San Jose, and Boston played its part in doing the dynamic world of high-performance computing proud. Our journey through GTC24 was one to be remembered, as we explored groundbreaking products, experienced the latest innovations from NVIDIA and formed connections with industry peers and partners.
At the heart of our presence at GTC24 was our booth, a bustling hub of activity where attendees gathered to witness firsthand the latest advancements in technology. Partnering with industry giants such as Supermicro and ZutaCore, we had an impressive array of products on display that offered a glimpse at the future of computing. From Supermicro's next-gen servers to ZutaCore's revolutionary cooling solutions, our booth was a testament to our commitment to pushing the boundaries of innovation.
We were also delighted to host SECQAI, adding a touch of quantum computing expertise to our exhibit. Their presence sparked engaging discussions and offered attendees an opportunity to peer into the transformative potential of quantum technology. There was some incredible tech on show throughout this year’s GTC, look at Boston’s stand outs, below!
The Blackwell is NVIDIA’s newest platform for building and deploying Trillion-Parameter Scale Generative AI, named in honour of David Harold Blackwell - a mathematician who specialised in game theory and statistics, and the first Black scholar inducted into the National Academy of Sciences.
Relative to its predecessor – Hopper, Blackwell achieves: 30X more AI Inference Performance, 4X faster AI Training, 25X lower energy use, and 25X lower TCO. It introduces groundbreaking advancements for accelerated computing through six technological innovations:
Blackwell GPUs are available in three system configurations: GB200 NVL72, HGX B200, and HGX B100 and will be available later this year.
For AI inference, the GB200 NVL72 achieves 30X more performance, 25X less energy, and 25X less cost than Hopper for trillion-parameter models. The HGX B200 achieves 15X more performance,12X less energy, and 12X less cost.
For AI training, the GB200 NVL72 achieves 4X more performance, 3.5X less energy, and 3.5X less cost than Hopper for trillion-parameter models. The HGX B200 achieves 3X more performance, 2.5X less energy, and 2.5X less cost.
For data processing, the GB200 NVL72 achieves 18X more performance, 7X less energy, and 5X less cost than CPUs for simulated database workloads. The HGX B200 achieves 6X more performance and 2X less energy.
It consists of 36 GB200 Grace Blackwell Superchips each with two Blackwell GPUs connected to one Grace CPU for a total of 72 Blackwell GPUs in a single NVLink domain. GB200 NVL72 is a multi-node, liquid-cooled, rack-scale system for the most compute-intensive workloads. The platform acts as a single GPU with 1.4 exaflops of AI performance and up to 30TB of fast memory and is a building block for the newly announced DGX SuperPOD. To achieve maximum compute density, the design uses liquid-cooling.
The premier accelerated x86 scale-up platforms designed for the most demanding generative AI, data analytics, and high-performance computing (HPC) workloads.
A liquid-cooled, rack-scale AI infrastructure based on the NVIDIA GB200 NVL72 platform that scales to thousands of NVIDIA GB200 Grace Blackwell Superchips for training and inferencing trillion parameter generative AI models.
This entirely new type of DGX SuperPOD features massive GPU memory and NVIDIA Grace CPUs, built on the ARM CPU architecture. It is a high compute density, liquid-cooled, rack-scale design. Racks are connected via NVIDIA Quantum-2 InfiniBand for DGX SuperPODs that can scale up to tens of thousands of GB200 Superchips. An optional configuration that connects 576 Blackwell GPUs with NVLink will also be available to power massive models.
The latest iteration of NVIDIA’s legendary DGX systems powered by NVIDIA Blackwell GPUs. DGX B200 is the proven choice for enterprises, a unified AI platform that accelerates any AI workload, from training, to fine-tuning, to inference.
The Quantum-X800 and Spectrum-X800 platforms are the next stage of InfiniBand and Ethernet evolution to support, respectively, the highest performance for AI, and optimised performance for AI cloud and enterprise.
The Quantum-X800 platform includes the NVIDIA Quantum Q3400 switch and NVIDIA ConnectX-8 SuperNIC to achieve end-to-end throughput of 800 Gb/s and provides SHARPv4 for 5x higher bandwidth capacity and 9x greater in-network computing compared to the previous generation.
The Spectrum-X800 platform includes the NVIDIA Spectrum SN5600 800 Gb/s switch and NVIDIA BlueField-3 SuperNIC optimised for multi-tenant AI cloud and enterprise infrastructure to facilitate tenant-isolation of AI workloads for optimal and consistent performance.
The experienced networking team at Boston are extremely excited about the launch of the X800 series and are keen to begin building out even more performant networks that are truly optimised for AI and HPC workloads.
NVIDIA AI Enterprise 5.0 is a major release update which introduces several new features and enhancements. Chief amongst these is the introduction NIM Microservices providing optimised runtime and building blocks for AI development – more on that later.
Alongside the introduction of NIM, support has been introduced for the following new GPUs:
As well as adding vGPU support for the following models which were previously bare metal only:
NIMS or NVIDIA Inference Micro Services to give it the full name is possibly one of the most interesting announcements made by Jenson at GTC ’24. It’s been revealed that a catalogue of NIMS will enable as many as two dozen popular pre-trained and optimised AI models will be provided either directly by NVIDIA or by its partner network via prebuilt containers.
Partners include,?A121, Adept,?Cohere, Getty Images, and Shutterstock as well as open models from Google,?Hugging Face, Meta, Microsoft, Mistral AI and Stability AI.
Customers with NVIDIA AI Enterprise 5 licenses will be able to access these as part of their existing license subscription, helping to minimise their time to take AI into production. NIMS provide industry-standard APIs for domains such as language, speech and drug discovery to enable developers to quickly build AI applications using their proprietary data hosted securely in their own infrastructure.
Naturally, Boston will be enabling and supporting our customers deploying NIMS on our wide range of NVIDIA-Certified systems, stay tuned or enquire to hear more about this exciting way to deploy AI models with ease.
Nvidia launched a new cloud-based platform - Nvidia Quantum Cloud, during Jenson’s GTC keynote. It’s designed to assist researchers, developers and organisations to access quantum computing clusters and quantum hardware through cloud-based platform which are fully managed and supplied by Nvidia, simplifying the whole process.
Nvidia also offers Quantum Development Kits which includes tutorials, documentation and example code to assist developers on their quantum computing journey. Using Nvidia Quantum Cloud, various use cases around quantum machine learning (such as quantum support vector machine and principal component analysis) and quantum neural networks (such as Image Classification) can be run.
Overall, Nvidia Quantum Cloud is a powerful platform for quantum computing and machine learning that can be used in a variety of applications, including cryptography, optimisation, and simulation.
Complimenting NVIDIA Quantum Cloud, here at Boston we have a similar approach for developing quantum computing projects in house. Our quantum centre of excellence, an extension of Boston Labs is hosted in state-of-the-art datacentre - Cyxtera LHR3. Similalrly, this lab is well equipped with computing resources and tools are designed for quantum computing in particular. It includes a wide range of quantum software tools and libraries such as Cirq, Qiskit and Pennylane which are super useful to execute quantum algorithms, optimise quantum circuits and quantum simulations.
This cloud-based platform is equipped with QPUs (Quantum Processing Units), GPU and CPU resources, meeting the needs of heterogenous computing. Boston Labs also includes dedicated platforms for quantum computing which are equipped with GPUs including RTX6000, L40S & H100, the latest generation CPUs from NVIDIA, AMD and Intel, pre-configured and optimised with all the required software dependencies.
Just like Nvidia Quantum Cloud, Boston Labs is designed to be accessible in user friendly fashion and open to all researchers, developers, and organisations of all sizes. Practitioners can experiment with quantum computing resources and hardware in very cost-effective ways and can avoid the need to invest in often expensive hardware and other infrastructure.
Paired with the Centre of Excellence, there is also Boston Training Academy, enabling education on Quantum and other computing paradigms for a wider audience. The academy covers tutorials around evolving quantum technology as well as traditional GPU and CPU computing methodologies.
To help our clients make informed decisions about new technologies, we have opened up our research & development facilities and actively encourage customers to try the latest platforms using their own tools and if necessary together with their existing hardware. Remote access is also available
The annual GPU Technology Conference is returning once again in San Jose, California!