Amazon Elastic Compute Cloud (Amazon EC2) accelerated computing portfolio affords the broadest alternative of accelerators to energy your synthetic intelligence (AI), machine studying (ML), graphics, and excessive efficiency computing (HPC) workloads. We’re excited to announce the growth of this portfolio with three new situations that includes the most recent NVIDIA GPUs: Amazon EC2 P5e situations powered by NVIDIA H200 GPUs, Amazon EC2 G6 situations that includes NVIDIA L4 GPUs, and Amazon EC2 G6e situations powered by NVIDIA L40S GPUs. All three situations can be accessible in 2024, and we look ahead to seeing what you are able to do with them.
AWS and NVIDIA have collaborated for over 13 years and have pioneered large-scale, extremely performant, and cost-effective GPU-based options for builders and enterprise throughout the spectrum. We’ve got mixed NVIDIA’s highly effective GPUs with differentiated AWS applied sciences corresponding to AWS Nitro System, 3,200 Gbps of Elastic Cloth Adapter (EFA) v2 networking, tons of of GB/s of information throughput with Amazon FSx for Lustre, and exascale computing with Amazon EC2 UltraClusters to ship probably the most performant infrastructure for AI/ML, graphics, and HPC. Coupled with different managed providers corresponding to Amazon Bedrock, Amazon SageMaker, and Amazon Elastic Kubernetes Service (Amazon EKS), these situations present builders with the trade’s finest platform for constructing and deploying generative AI, HPC, and graphics functions.
Excessive-performance and cost-effective GPU-based situations for AI, HPC, and graphics workloads
To energy the event, coaching, and inference of the most important giant language fashions (LLMs), EC2 P5e situations will function NVIDIA’s newest H200 GPUs, which provide 141 GBs of HBM3e GPU reminiscence, which is 1.7 occasions bigger and 1.4 occasions quicker than H100 GPUs. This enhance in GPU reminiscence together with as much as 3200 Gbps of EFA networking enabled by AWS Nitro System will allow you to proceed to construct, prepare, and deploy your cutting-edge fashions on AWS.
EC2 G6e situations, that includes NVIDIA L40S GPUs, are constructed to supply builders with a broadly accessible possibility for coaching and inference of publicly accessible LLMs, in addition to help the rising adoption of Small Language Fashions (SLM). They’re additionally optimum for digital twin functions that use NVIDIA Omniverse for describing and simulating throughout 3D instruments and functions, and for creating digital worlds and superior workflows for industrial digitalization.
EC2 G6 situations, that includes NVIDIA L4 GPUs, will ship a lower-cost, energy-efficient answer for deploying ML fashions for pure language processing, language translation, video and picture evaluation, speech recognition, and personalization in addition to graphics workloads, corresponding to creating and rendering real-time, cinematic-quality graphics and sport streaming.
Concerning the Creator
Chetan Kapoor is the Director of Product Administration for the Amazon EC2 Accelerated Computing Portfolio.