Partner Spotlight - NVIDIA 

NVidia

NVIDIA (Nasdaq: NVDA) is the world leader in visual computing technologies and the inventor of the GPU, a high-performance processor which generates breathtaking, interactive graphics on workstations, personal computers, game consoles, and mobile devices.

gpucomputing
The most advanced data centre GPU ever

Imagine a processor that offers advanced design capabilities for entertainment and commerce, while enabling scientists to tackle the world's most challenging problems from climate research to quantum physics. The GPU makes all of this possible.

GPU computing is the use of a GPU (graphics processing unit) together with a CPU to accelerate general-purpose scientific and engineering applications. Introduced by NVIDIA, GPU computing has quickly become an industry standard, enjoyed by millions of users worldwide and adopted by virtually all computing vendors.

GPU computing offers unprecedented application performance by offloading compute-intensive portions of the application to the GPU, while the remainder of the code still runs on the CPU. From a user's perspective, applications simply run significantly faster.

CPU vs. GPU

A simple way to understand the difference between a CPU and GPU is to compare how they process tasks. A CPU consists of a few cores that are optimised for sequential serial processing, whereas a GPU consists of thousands of smaller, more efficient cores designed for handling multiple tasks simultaneously.

NVIDIA Tesla 2
The most advanced data centre GPU ever

The most advanced data centre GPU ever

microsite-p100
The most advanced data centre GPU ever

Artificial intelligence for self-driving cars. Predicting our climate's future. A new drug to treat cancer. Some of the world's most important challenges need to be solved today, but require tremendous amounts of computing to become reality. Today's large-scale data center relies on lots of interconnected commodity compute nodes, limiting the performance needed to drive these important workloads. Now, more than ever, the data center must prepare for the high-performance computing and hyperscale workloads being thrust upon it.

The NVIDIA® Tesla® P100 is purpose-built as the most advanced data center accelerator ever. It taps into an innovative new GPU architecture to deliver the world’s fastest compute node capable of performance equal to hundreds of slower commodity compute nodes. Lightning-fast nodes powered by Tesla P100 accelerate time-to-solution for the world’s most important challenges that have infinite compute needs in HPC and deep learning.

P100 features five ground breaking technologies:

  • New Pascal Architecture: Delivering 5.3 and 10.6 TeraFLOPS of double and single precision for HPC, 21.2 TeraFLOPS of FP16 for Deep learning
  • NVLink: World's first high-speed Interconnect for multi-GPU scalability with 5x boost in performance
  • CoWoS ® with HBM2: Unifying data and compute into single package for up to 3X memory bandwidth over prior-generation solution
  • 16nm FinFET for unprecedented energy efficiency - With 15.3 billion transistors built on 16 nanometer FinFET fabrication technology, the Pascal GPU is the world's largest FinFET chip ever built2. It is engineered to deliver the fastest performance and best energy efficiency for workloads with near-infinite computing needs.
  • New AI Algorithms: New half-precision, 16-bit floating point instructions deliver over 21 TeraFLOPS for unprecedented training performance.
  P100 for NVLink-optimized Servers P100 for PCIe-based Servers
Double-Precision Performance 5.3 TeraFLOPS 4.7 TeraFLOPS
Single-Precision Performance  10.6 TeraFLOPS 9.3 TeraFLOPS
Half-Precision Performance 21.2 TeraFLOPS 18.7 TeraFLOPS
NVIDIA NVLink™ Interconnect Bandwidth 160GB/s -
PCIe x16 Interconnect Bandwidth 32GB/s 32GB/s
CoWoS HBM2 Stacked Memory Capacity 16GB 16GB 12GB
CoWoS HBM2 Stacked Memory Bandwidth 720GB/s 720GB/s 540GB/s
Enhanced Programability with Page Migration Engine Yes Yes
ECC Protection for Reliability Yes Yes
Server-Optimized for Data Center Deployment Yes Yes

* FLOPS performance with NVIDIA GPU Boost™ * Interconnect Bandwidth measured Bidirectional

Boost up with NVIDIA Tesla K80 GPU accelerators

NVIDIA Tesla 3
Boost up with NVIDIA Tesla K80 GPU accelerators

The Tesla K80 accelerator is a dual GPU board designed to cope with the most demanding computational workloads. It has 24GB of GPU accelerator memory meaning that the Tesla K80 delivers a 10x speed-up compared to the latest CPUs and TWICE the performance of its predecessor, the Tesla K20.

The K80 is based on the NVIDIA Kepler Architecture and the widely used CUDA® parallel computing model. The Tesla K-series GPUs are the leading platform for speeding through big data analytics and scientific computing workloads, which has a widespread ecosystem of software developers, software vendors and datacentre OEMs to accelerate findings and insight.

NVIDIA Tesla 4
CPU comparisons

The Boston ANNA Pascal

microsite-teslaboost

Introducing the Boston ANNA Pascal GPU server - with support for up to 4x NVIDIA Tesla P100 (codename Pascal) GPUs.

The revolutionary NVIDIA® Pascal™ architecture is purpose-built to be the engine of computers that learn, see, and simulate our world - making it the perfect GPU backbone for the 1028GQ-TXR and ensuring it is a contender for the title of 'world's fastest deep learning appliance'.

Powered by four ground-breaking technologies with discontinuous jumps in performance, Tesla P100 enables lightning fast nodes to deliver the highest absolute performance for HPC and deep learning workloads with infinite computing needs.

One of the most interesting features of Pascal is the new NVLink interconnect system. When using a standard multi-GPU configuration with a PCI-E switch, the limited PCI-E bandwidth to the CPU memory is shared between GPU's. The link will become even more saturated when peer-topeer GPU traffic is factored in. NVLink addresses this problem by providing a high-speed 80Gb/s interconnect path wholly devoted to peer GPU-to-GPU connections. Combining this with RDMA via Infiniband or Omnipath provides the most powerful possible parallel computing environment. As shown below with the increase in performance compared to the previous generation K80 GPUs.

Boston's industry leading server is a masterclass of server design and innovation, creating a streamlined architecture which eliminates complex cabling and GPU pre-heat for maximum airflow, cooling and performance-per-watt. Housing 3 GPUs in the front and 1 at the rear, no PCI-E extension cables or repeaters are used, instead GPU's are connected via the PCI-E Gen 3 bus providing reduced latency. In addition to the 4 GPUs, the 1U server provides 2 x PCI-E Gen 3 slots for InfiniBand or Omnipath allowing strong RDMA performance.

annapascal

Key Features

  • NVIDIA Pascal Architecture
  • NVLINK
  • High Density
  • Peer-to-Peer GPU communication

If you are interested in testing one of our these solutions, please let us know by completing the short form below. Once submitted, one of our team will contact you to discuss your requirements and the options available.

*
*
*
Enter the text shown in the image:

Find your solution

Test out any of our solutions at Boston Labs

To help our clients make informed decisions about new technologies, we have opened up our research & development facilities and actively encourage customers to try the latest platforms using their own tools and if necessary together with their existing hardware. Remote access is also available

Contact us

whd2017

Latest Event

WHD.global|25th-31st March 2017, Europa-Park, Rust, Germany

WorldHostingDays is the biggest event of its kind for the cloud services and internet infrastructure industries. WHD.global is a week-long festival taking place in Germany annually. More than 6,500 attendees can be expected this year who will gain a clearer picture of cloud services and internet infrastructure.

More Info