Each of the five servers will address multiple AI computing scenarios and support 8 to 16 of the latest NVIDIA A100 Tensor Core GPUs. The third generation of Tensor Cores in A100 GPUs are faster, more flexible and easier to use and enable these servers to offer AI computing performance of up to 40 PetaOPS. With next-gen NVIDIA NVSwitch fabric technology, the GPU-to-GPU P2P communication speed has been doubled to reach 600 GB/sec. The increased computing performance and updated GPU-to-GPU bandwidth will bring AI computing efficiency to enable AI training for larger data scales and more complicated models. In addition, the NVIDIA A100 GPUs have 40GB of HBM2 memory with 70% higher memory bandwidth at 1,6TB/sec to support larger deep learning models for training.
Every AI server will offer hardware design for several application scenarios, including:
Inspur will also upgrade its leading AI computing resource platform AIStation and automatic machine learning platform AutoML Suite to support NVIDIA A100, which provide more flexible AI computing system resources management and powerful AI modeling algorithm development support.
Additionally, Inspur plans to add the new EGX A100 configuration to its edge server portfolio to deliver enhanced security and unprecedented performance at the edge. The EGX A100 converged accelerator combines an NVIDIA Mellanox SmartNIC with GPUs powered by the new NVIDIA Ampere architecture, so enterprises can run AI at the edge more securely.
"Inspur was swift to unveil upgrades to its AI servers based on the NVIDIA Ampere architecture as it allows global AI users a broad array of computing platforms optimized for various applications", stated Liu Jun, GM of AI and HPC at Inspur. "With this upgrade, Inspur offers the most comprehensive AI server portfolio in the industry, better tackling the computing challenges created by data surges and complex modelling. We expect that the upgrade will significantly boost AI technology innovation and applications."
"NVIDIA A100 Tensor Core GPUs offer customers unmatched acceleration at every scale for AI, data analytics and HPC", stated Paresh Kharya, Director of Product Management for Accelerated Computing at NVIDIA. "Inspur AI servers, powered by NVIDIA A100 GPUs, will help global users eliminate their computing bottlenecks and dramatically lower their cost, energy consumption, and data centre space requirements."
Inspur, a global AI server provider, offers a comprehensive AI product portfolio. It also fosters close partnerships with leading AI customers and empowers them to enhance performance in AI applications like voice, image, video, language and search.