Nvidia introduced a new Jump Start program for its DGX Station that lowers the introduction pricing by 25% for the first system you purchase. Even amid the soul-crushing GPU shortage, $49,900 may seem like a steep price to pay for a system that comes wielding four GPUs, but these aren't garden-variety cards. The workstation features Nvidia's Volta V100 GPUs, which are designed for professional applications. The DGX Station, in particular, is designed specifically for deep learning developers.
Nvidia designed its DGX Station to be a smaller workstation-class version of its DGX-1 server. The four watercooled Tesla V100s churn out up to 500 TeraFLOPS of performance inside the full-tower design. That comes courtesy of 64GB of HBM2 memory and a total of 20,480 CUDA cores.
The DGX Station comes loaded with a range of other performance-boosting additives, such as 256GB of DDR4 memory and three 1.92TB SSDs in RAID 0. It also includes a 1.92TB SSD for the operating system and a dual 10Gb LAN networking solution.
Nvidia claimed the system can outperform 400 CPUs (no mention of which CPUs) with a mere fraction of the power consumption. Granted, this system still consumes 1,500W under full load, but Nvidia claimed the water cooling ensures that it's quiet enough to set under your desk during operation. That means deep learning developers can use the system in their office, which often isn't possible with bulkier servers.
The price tag may seem extreme, but the system has exclusive capabilities. For now, it is the only workstation available that can run four V100s together using NVLink, whereas other workstations, including the ones individual devs build on their own, are limited to dual-NVLink implementations. The NVLink connection boosts performance with broader data pipes between each GPU. For instance, the four-way NVLink confers a 5x inter-GPU bandwidth benefit over a standard PCIe connection between the GPUs.
Nvidia also includes its software suite, which comes pre-qualified and loaded onto the system, which the company claims offers up to a 30% performance advantage. Developers can simply load pre-optimized containers, design their software, and then migrate it to production-class servers in remote data centers. Of course, Nvidia imagines these systems working in tandem with its DGX-1 servers (pictured above), which allows the developer to migrate the code without recompiling.
Frank Wu, the Director of Machine Learning at SAP, explained that one of the immediate benefits of the DGX station was a drastic reduction in the time it took to train deep learning models. Previously, the company had used a CPU-based system in its lab, but the DGX Station's compact and silent design allowed the company to use it in a typical office setting. The company designs the models in the office, then migrates them directly from the DGX Station to the DGX-1 server in a remote data center. SAP also found that the system has enough horsepower to support multiple simultaneous users, which is ideal for its deep learning group because the system is shared with teams in both the U.S. and in Germany. For instance, SAP's HANA group also uses the system to develop its GPU-accelerated database software.
The DGX Station's 25% discount comes via resellers, and it applies only to the first system purchase. Any follow-on purchases come with the normal $69,000 price tag. The Jump-Start program runs until April 29, 2018.