This week, Google Cloud Next is taking place in Las Vegas, bringing forth a wave of new instance types and accelerators to enhance the Google Cloud Platform. Amidst the unveiling of custom Arm-based Axion chips, the spotlight of this year’s announcements primarily shines on AI accelerators, whether crafted by Google or in collaboration with Nvidia.
Before Blackwell: Pioneering Innovations
Before the public availability of Nvidia’s Blackwell chips, Google steps forward with proactive solutions. For developers seeking enhanced capabilities to train Large Language Models (LLMs) promptly, Google introduces the A3 Mega instance. This collaborative effort with Nvidia integrates industry-standard H100 GPUs with a novel networking system, capable of delivering up to twice the bandwidth per GPU.
Another noteworthy addition is the A3 Confidential instance, designed to fortify the confidentiality and integrity of sensitive data and AI workloads during both training and inferencing processes. Leveraging confidential computing, data transfers between Intel’s CPU and Nvidia’s H100 GPU are encrypted via protected PCIe, ensuring heightened security without necessitating code modifications.
In tandem with third-party offerings, Google unveils its in-house innovation with the launch of Cloud TPU v5p processors — the most robust AI accelerators to date. Boasting a twofold enhancement in floating point operations per second and a threefold increase in memory bandwidth speed, these chips signify a significant leap forward in AI computation capabilities.
Amidst the buzz surrounding fast-paced innovations, Google acknowledges the importance of a complementary underlying architecture to harness the full potential of these advancements. In response, Google introduces new AI-optimized storage options, including the preview release of Hyperdisk ML. This next-generation block storage service promises to accelerate model load times by up to 3.7x, aligning storage capabilities with the rapid pace of AI computation.
Expanding the Horizon of Compute Capabilities
Beyond the realm of AI acceleration, Google broadens its compute offerings with a range of traditional instances, powered by Intel’s latest Xeon processors. Among these, the newly unveiled C4 and N4 instances stand out, featuring the fifth-generation Emerald Rapids Xeons. While the C4 instance prioritizes performance, the N4 instance offers a cost-effective solution, catering to diverse workload requirements.
Additionally, Google introduces innovative bare-metal machines, including the C3 series powered by fourth-generation Intel Xeons, and the X4 memory-optimized bare metal instances, both currently in preview. Furthermore, the introduction of the Z3 instance marks Google Cloud’s foray into storage-optimized virtual machines, promising unparalleled IOPS for storage-intensive workloads among leading cloud providers.
In summary, Google Cloud Next unveils a plethora of cutting-edge solutions, spanning AI acceleration, storage optimization, and compute capabilities, reaffirming Google’s commitment to driving innovation in cloud computing.
See also: Google Vids: Empowering Personalized Video Creation With AI Assistance