NVIDIA A100 PCIe: Accelerator with 6912 CUDA Cores and 250W TDP

Although we usually talk about NVIDIA as a company dedicated to gaming hardware, the company is much more. Above all, it is very well valued within the advanced computing and Artificial Intelligence market. To meet the needs in these two fields, the company has just announced its NVIDIA A100 PCIe card. This new accelerator promises great performance with a tight consumption.

This new NVIDIA A100 PCIe, as the company states, is intended for "AI, data science and scientific computing." It is also characterized by being based on the Ampere architecture and therefore, the 7nm lithograph. Additionally, this graphic is based on PCI-Express 4.0, so it has been designed to work with AMD EPYC processors.

Gigabyte Technology GeForce RTX 2060 OC GV-N2060OC-6GD - Graphics Card, Black
  • Powered by geforce rtx 2060
  • Integrated with 6GB GDDR6 192-bit memory interface
  • Windforce 2x cooling system with alternate fans
  • Unique 90mm blade fans

NVIDIA A100 PCIe, a high power and low consumption accelerator

The new NVIDIA A100 PCIe in specifications is very similar to the NVIDIA A100 SXM model, but with a reduced TDP. It goes from a TDP of 400W for the SXM model to a TDP of 250W for this A100 PCIe model.

Said NVIDIA solution features an 826mm ID2 with no less than 54.000 million transistors. This solution integrates 6912 CUDA Cores and 432 Tensor Cores at a frequency of 1.410MHz. It doesn't use full silicon with all the CUDA Cores, so we could see other versions even more powerful. This solution has a FP32 power of 19.5 TFLOPs, an FP64 performance of 9.7 TFLOPs and a FP4 Tensor Cores performance of 19.5 TFLOPs.

Nvidia-A100-PCIe-graphics

Additionally, this NVIDIA A100 PCIe has 40GB HBM2E with a memory interface of 5.120bits with a bandwidth of 1.555GB / s. This solution makes use of a third generation NVLink system for a bandwidth of 4.8TB / s. Additionally, this solution is based on the PCI-Express 4.0 / SXM4 interface.

Because it is designed for Artificial Intelligence, this one has new Tensor Cores twice as fast as the previous generation. Additionally, it can be up to x20 times faster than the Tesla V100 for Artificial Intelligence.

Press release

NVIDIA and the world's leading server manufacturers have just introduced their systems based on NVIDIA A100 technology with a wide variety of designs and configurations to meet the most complex challenges of AI, data science and scientific computing.

Following the release of the NVIDIA Ampere architecture and NVIDIA A100 GPU last month, more than 50 A100 servers from the world's leading manufacturers are expected to be installed, including Asus, Atos, Cisco, Dell Technologies, Fujitsu. , Gigabyte, Hewlett Packard Enterprise, Inspur, Lenovo, One Stop Systems, Quanta / QCT, and Supermicro.

Server availability varies, with 30 systems planned this summer, and more than 20 more by the end of the year.

“Adoption of NVIDIA A100 GPUs across leading server vendors’ offerings is outpacing anything we’ve seen before,” said Ian Buck, vice president and general manager of accelerated computing at NVIDIA. “The sheer breadth of NVIDIA A100 servers from our partners ensures that customers can choose the best options to accelerate their data centers for high utilization and low total cost of ownership.”

The first GPU based on NVIDIA's Ampere architecture, the A100 can boost performance up to 20 times more than its predecessor, making it the company's biggest GPU performance leap to date. It includes several technical advancements, including a new multi-instance GPU technology that allows a single A100 to be split into up to seven different GPUs to handle different computing tasks; XNUMXrd generation NVIDIA NVLink technology that allows multiple GPUs to be joined together to function as a single giant GPU; and new structural spreading features that can be used to double the performance of a GPU.

Exit mobile version