"Deep learning, machine learning and artificial intelligence are vital tools for addressing the world's most complex challenges and improving many aspects of our lives", stated William Wu, Director of Product Management, Penguin Computing. "Our breadth of products covers configurations that accelerate various demanding workloads - maximizing performance, minimizing P2P latency of multiple GPUs and providing minimal power consumption through creative cooling solutions."
NVIDIA Tesla V100 GPUs join an expansive GPU server line that covers Penguin Computing's Relion servers (Intel-based) and Altus servers (AMD-based) in both 19" and 21" Tundra form factors. Penguin Computing will debut a high density 21" Tundra 1OU GPU server to support 4x Tesla V100 SXM2, and 19" 4U GPU server to support 8x Tesla V100 SXM2 with NVIDIA NVLink interconnect technology optional in single root complex.
The NVIDIA Volta architecture is bolstered by pairing NVIDIA CUDA cores and NVIDIA Tensor Cores within a unified architecture. A single server with Tesla V100 GPUs can replace hundreds of CPU servers for AI. Equipped with 640 Tensor Cores, Tesla V100 delivers 125 TeraFLOPS of deep learning performance. That's 12X Tensor FLOPS for deep learning training, and 6X Tensor FLOPS for deep learning inference when compared to NVIDIA Pascal GPUs.
"Penguin Computing continues to demonstrate leadership by providing Volta-based systems to support critical AI research", stated Paresh Kharya, Group Product Marketing Manager, NVIDIA. "Tesla V100 systems will enable their customers to create innovative AI products and services by accelerating their AI research and deployments."
The announcement reinforces Penguin Computing's philosophy and broader capabilities as a full-spectrum provider offering complete solutions. This includes tailored, custom designs that are supportable and scale to large deployments, and fully engineered and architected designs.