AWS Announces Three New Amazon EC2 Instances Powered by AWS-Designed Chips

Amazon EC2 C7g instances powered by new AWS Graviton3 processors provide up to 25% better performance for compute-intensive workloads over current generation C6g instances powered by AWS Graviton2 processors Amazon EC2 Trn1 instances powered by AWS Trainium chips provide the best price performance and the fastest time to train most machine learning models in Amazon EC2 Amazon EC2 Im4gn/Is4gen/I4i instances feature new AWS Nitro SSDs for the best storage performance for I/O-intensive workloads

Amazon Web Services (AWS) announced three new Amazon Elastic Compute Cloud  (Amazon EC2) instances powered by AWS-designed chips that help customers significantly improve the  performance, cost, and energy efficiency of their workloads running on Amazon EC2. New C7g instances  powered by next-generation AWS Graviton3 processors deliver up to 25% better performance  than current generation C6g instances powered by AWS Graviton2 processors. 

New Trn1 instances  powered by AWS Trainium chips provide the best price performance and the fastest time to train most  machine learning models in Amazon EC2. New storage-optimized Im4gn/Is4gen/I4i instances based on AWS-designed AWS Nitro SSDs (solid-state drives) offer the best storage performance for I/O-intensive  workloads running on Amazon EC2. Together, these instances herald the arrival of new Amazon EC2  instances based on AWS-designed chips that help customers power their most business-critical applications. 

“With our investments in AWS-designed chips, customers have realized huge price performance benefits for some of today’s most business-critical workloads. These customers have asked us to continue  pushing the envelope with each new EC2 instance generation,” said David Brown, Vice President,  Amazon EC2 at AWS. “AWS’s continued innovation means customers are now getting brand new, game  changing instances to run their most important workloads with significantly better price performance  than anywhere else.” 

C7g instances powered by new AWS Graviton3 processors deliver up to 25% better performance  compared to current generation C6g instances powered by AWS Graviton2 processors Customers like DirecTV, Discovery, Epic Games, Formula 1, Honeycomb.io, Intuit, Lyft, MercardoLibre,  NextRoll, Nielsen, SmugMug, Snap, Splunk, and Sprinklr have seen significant performance gains and  reduced costs from running AWS Graviton2-based instances in production since they launched in 2020.  The Graviton2 instance portfolio offers 12 different instances that include general purpose, compute  optimized, memory optimized, storage optimized, burstable, and accelerated computing instances, so  customers have the deepest and broadest choice of cost-effective and power-efficient compute in the  cloud. As customers bring more compute intensive workloads like high performance computing (HPC),  gaming, and machine learning inference to the cloud, and as their compute, storage, memory, and  networking demands grow, they are looking for even better price performance and energy efficiency to  run these demanding workloads.  

C7g instances, powered by next generation AWS Graviton3 processors, provide up to 25% better performance for compute-intensive workloads compared to current generation C6g instances powered  by Graviton2 processors. AWS Graviton3 processors also deliver up to 2x higher floating point performance for scientific, machine learning, and media encoding workloads, up to 2x faster  performance for cryptographic workloads, and up to 3x better performance for machine learning  workloads compared to previous generation AWS Graviton2 processors.

AWS Graviton3 processors are  also more energy efficient, using up to 60% less energy for same performance than comparable EC2  instances. C7g instances are the first in the cloud to feature the latest DDR5 memory, which provides  50% higher memory bandwidth versus AWS Graviton2-based instances to improve the performance of  memory-intensive applications like scientific computing. C7g instances also deliver 20% higher  networking bandwidth compared to AWS Graviton2-based instances.

C7g instances support Elastic  Fabric Adapter (EFA), which allows applications to communicate directly with network interface cards, providing lower and more consistent latency, to enhance the performance of applications that require  parallel processing at scale like HPC and video encoding. C7g instances are available today in preview.

Trn1 instances powered by AWS Trainium chips provide the best price performance and the fastest  time to train most machine learning models in Amazon EC2 

More and more customers are building, training, and deploying machine learning models to power  applications that have the potential to reinvent their businesses and customer experiences. However, to  ensure improved accuracy, these machine learning models must consume ever-growing amounts of  training data, which causes them to become increasingly expensive to train. This dilemma can have the  effect of limiting the number of machine learning models that customers are able to deploy. AWS provides the broadest and deepest choice of compute offerings for machine learning, including the EC2  P4d instances featuring NVIDIA A100 Tensor Core GPUs and EC2 DL1 instances featuring Gaudi  accelerators from Habana Labs. But even with the fastest accelerated instances available today, it can  still be prohibitively expensive and time consuming to train ever-larger machine learning models.  

Trn1 instances powered by AWS Trainium chips offer the best price performance and the fastest  machine learning model training in Amazon EC2, providing up to 40% lower cost to train deep learning  models compared to the latest P4d instances. Trn1 instances offer 800 Gbps EFA networking bandwidth  (2x higher than the latest EC2 GPU-based instances) and integrate with Amazon FSx for Lustre high  performance storage—enabling customers to launch Trn1 instances with EC2 UltraClusters capability. 

With EC2 UltraClusters, developers can scale machine learning training to 10,000+ Trainium accelerators  interconnected with petabit-scale networking, giving customers on-demand access to supercomputing class performance to cut training time from months to days for even the largest and most complex  models. Trn1 instances are available today in preview.

Im4gn/Is4gen/I4i instances featuring new AWS Nitro SSDs deliver the best storage performance for  I/O intensive-workloads 

Today, customers use I3/I3en storage-optimized instances for applications that require direct access to  data sets on local storage like scale-out transactional and relational databases (e.g. MySQL and  PostgreSQL), NoSQL databases (e.g. Cassandra, MongoDB, Redis, etc.), big data (e.g. Hadoop), and data  analytics workloads (e.g. Spark, Hive, Presto, etc.). I3/I3en instances offer Non-Volatile Memory Express  (NVMe) SSD-backed instance storage optimized for low latency, high I/O performance, and throughput  at a low cost. Customers appreciate the fast transaction times I3/I3en instances provide, but as they  evolve their workloads to process even more complex transactions on larger data sets, they need even  higher compute performance and faster access to data, without higher costs.

Im4gn/Is4gen/I4i instances are architected to maximize the storage performance of I/O-intensive  workloads. Im4gn/Is4gen/I4i instances offer up to 30 TB of NVMe storage from AWS-designed AWS  Nitro SSDs, delivering up to 60% lower I/O latency and 75% lower latency variability compared to previous generation I3 instances to maximize application performance.

AWS Nitro SSDs are tightly  integrated with the AWS Nitro System via optimizations in the storage stack, hypervisor, and hardware. Because AWS is managing both the hardware and firmware of the AWS Nitro SSDs, customers benefit  from improved functionality because SSD updates are delivered more quickly compared to using  commercial SSDs. Im4gn instances (available today) feature AWS Graviton2 processors and provide up  to 40% better price performance and up to 44% lower cost per TB of storage compared to I3 instances. 

Is4gen instances (available today) also use AWS Graviton2 processors and provide up to 15% lower cost  per TB of storage and up to 48% better compute performance compared to I3en instances. I4i instances  (available soon) feature 3rd generation Intel Scalable processors (Ice Lake), delivering up to 55% better  compute performance than current generation I3 instances.

SAP HANA is a world’s leading in-memory database that serves as the foundation of the SAP Business  Technology Platform. “Over the past decade, SAP HANA has helped customers manage their most  mission critical transactional and analytics workloads,” said Irfan Khan, President of HANA Database &  Analytics at SAP. “AWS investments and innovations on ARM-based AWS Graviton processors and SAP  HANA Cloud are a great match with potential to deliver step-wise operation and performance  improvement benefits to our enterprise customers, and to SAP’s cloud analytics and data management  solutions powered by SAP HANA Cloud.” 

Twitter is what’s happening and what people are talking about right now. “Twitter is working on a multi year project to leverage the AWS Graviton-based EC2 instances to deliver Twitter timelines. As part of  our ongoing engineering to drive further efficiencies, we tested the new Graviton3-based C7g  instances,” said Nick Tornow, Head of Platform at Twitter. “Across a number of benchmarks that we’ve  found to be representative of the performance of Twitter workloads, we found Graviton3-based C7g  instances deliver 20%-80% higher performance versus Graviton2-based C6g instances, while also  reducing tail latencies by as much as 35%. We are excited to utilize Graviton3-based instances in the  future to realize significant price performance benefits.”

Formula 1 (F1) racing began in 1950 and is the world’s most prestigious motor racing competition, as  well as the world’s most popular annual sporting series. “We had already seen that Graviton2-based  C6gn instances provided us the best price performance for some of our CFD workloads. We have now  found Graviton3 C7g instances to be 40% faster than the Graviton2 C6gn instances for those same  simulations,” said Pat Symonds, CTO at Formula 1 Management. “We’re excited that EFA will be  standard on this instance type, and given this much improved price performance, we expect Graviton3- based instances to become the optimal choice to run all of our CFD workloads.” 

Founded in 1991, Epic Games is the creator of Fortnite, Unreal, Gears of War, Shadow Complex, and the  Infinity Blade series of games. Epic’s Unreal Engine technology brings high-fidelity, interactive  experiences to PC, console, mobile, AR, VR, and the Web. “As we look to the future and building  increasingly immersive and compelling experiences for players, we are excited to use AWS Graviton3- based EC2 instances,” said Mark Imbriaco, Senior Director of Engineering at Epic Games. “Our testing  has shown they are suitable for even the most demanding, latency-sensitive workloads while providing significant price performance benefits and expanding what is possible within Fortnite and any Unreal  Engine created experience.” 

Honeycomb develops an observability platform that enables engineering teams to visualize, analyze,  and improve cloud application quality and performance. “We’re excited to have tested our high throughput telemetry ingestion workload against early preview instances of AWS Graviton3 and have  seen a 35% performance increase for our workload over Graviton2,” said Liz Fong-Jones, Principal  Developer Advocate at honeycomb.io. “We were able to run 30% fewer instances of C7g than C6g  serving the same workload, and with 30% reduced latency. We are looking forward to adopting AWS  Graviton3-powered C7g instances in production once they are generally available.” 

Anthropic builds reliable, interpretable, and steerable AI systems that will have many opportunities to  create value commercially and for public benefit. “Our research interests span multiple areas including  natural language, human feedback, scaling laws, reinforcement learning, code generation, and  interpretability. A major key to our success is access to modern infrastructure that allows us to spin up  very large fleets of high-performance deep learning accelerators,” said Tom Brown, Co-founder at  Anthropic. “We are looking forward to using Trn1 instances powered by AWS Trainium, as their  unprecedented ability to scale to tens of thousands of nodes and higher network bandwidth will enable  us to iterate faster while keeping our costs under control.”

Splunk is a leading data platform provider and is designed to investigate, monitor, analyze, and act on  data at any scale. “We run C/C++ based workloads for indexing and searching event data. Our workload  is CPU bound and benefits from high capacity and low latency SSD storage,” said Brad Murphy, Vice  President, Cloud Platform & Infrastructure at Splunk. “When evaluating the new Im4gn/Is4gen instances  powered by AWS Graviton2, we observed an up to 50% decrease in search runtime compared to I3/I3en  instances, which we currently use. This makes Im4gn and Is4gen instances a great choice for running our  storage-intensive workloads with significant price performance improvement and lower TCO.” 

Sprinklr helps the world’s biggest companies make their customers happier across 30+ digital channels— using the most advanced, sophisticated AI engine built for the enterprise to create insight-driven  strategies and better customer experiences. “We benchmarked our Java-based search workloads on  Amazon EC2 Im4gn/Is4gen instances powered by AWS Graviton2 processors. Smaller Is4gen instances  offer similar performance compared to larger I3en instances, presenting an opportunity to meaningfully  reduce the TCO,” said Abhay Bansal, Vice President of Engineering at Sprinklr. “We also saw a significant  50% reduction in latency for queries when moving our workloads from I3 to Im4gn instances, indicating  a significant 40% price performance benefit. Moving to AWS Graviton2-based instances was easy, taking  two weeks to complete benchmarking. We are very happy with our experience and look forward to  running these workloads in production on Im4gn and Is4gen instances.”  

Redis Enterprise powers mission critical apps and services for over 8,000 organizations globally by  enabling software teams to create a high-performance data layer for the real-time world. “We’re thrilled  to see the Amazon EC2 I4i instances using the new low latency AWS Nitro SSDs that deliver better  transaction speed than the previous generation instances,” said Yiftach Shoolman, Co-Founder and CTO  at Redis. “We expect the faster storage performance and higher networking and processor speeds of the  I4i instances will deliver significant improvements at an even more attractive total cost of ownership for  our customers who use Redis-on-Flash on I4i instances.”