Google Cloud adds Nvidia P4 graphics cards for AI and virtual desktops
Google LLC today made a new graphics accelerator available in its public cloud to provide better support for artificial intelligence and virtual desktop workloads.
The chip in question is market leader Nvidia Corp.’s P4. It brings the number of Nvidia graphics processing units that Google’s cloud platform supports to four, all of which have been added since February 2017. The pace at which the company has expanded its GPU lineup reflects just how fast enterprise AI adoption is growing.
With a starting price of 60 cents per hour, the P4 is the second most affordable of the four available GPUs. The chip can provide 5.5 teraflops of performance when processing single-precision values that take up 4 bytes. A single teraflop equals 1 trillion floating-point operations per second, a standard unit of computing power.
Nvidia has also equipped the P4 with 8 gigabytes of GDDR5 memory, which is specifically designed for use in GPUs. On-chip memory is faster than the regular standalone kind because it keeps data closer to the GPU cores and thereby cuts latency.
In AI deployments, Google sees its new cloud-based P4s being used mainly for machine learning inference. That’s the data processing neural networks do when they’re in production, after they have been properly trained, which is an entirely different task that’s sometimes better served by more powerful GPUs.
The P4 lends itself just as well to virtual desktop environments. It features Grid, a piece of software from Nvidia that provides pass-through access to GPU resources for virtual machines. For added measure, Google offers access to a tool from partner Teradici Inc. that can stream footage of applications running inside a VM to an employee’s local device.
The third and final use case the company is targeting with its cloud-based P4s is video streaming. The chip has three video processing engines that, according to Nvidia, can transcode up to 35 high-definition streams in real time.
Nvidia is a key partner for Google’s efforts to address the growing role of GPUs in companies’ technology strategies. With that said, Google is not fully reliant on the chipmaker for AI processors. The company also offers cloud customers its Tensor Processing Units, internally designed chips customized for running neural networks that can each provide a massive 180 teraflops of computing power.
Image: Google
Since you’re here …
… We’d like to tell you about our mission and how you can help us fulfill it. SiliconANGLE Media Inc.’s business model is based on the intrinsic value of the content, not advertising. Unlike many online publications, we don’t have a paywall or run banner advertising, because we want to keep our journalism open, without influence or the need to chase traffic.The journalism, reporting and commentary on SiliconANGLE — along with live, unscripted video from our Silicon Valley studio and globe-trotting video teams at theCUBE — take a lot of hard work, time and money. Keeping the quality high requires the support of sponsors who are aligned with our vision of ad-free journalism content.
If you like the reporting, video interviews and other ad-free content here, please take a moment to check out a sample of the video content supported by our sponsors, tweet your support, and keep coming back to SiliconANGLE.