Uber extends its agreement with AWS and tests Trainium3 for its AI engine

Uber has decided to strengthen its relationship with Amazon Web Services with a move that goes beyond mere cloud capacity. The company has expanded its use of AWS infrastructure to handle more load for its real-time matching systems on Graviton4 instances and has also begun pilot testing the training of some AI models on Trainium3, AWS’s chips specifically designed for AI training. This development is significant because it shows how Uber is fine-tuning its multi-cloud strategy not only based on the provider but also on the type of silicon that best fits each workload.

According to AWS, Uber will utilize more Graviton4 instances to support its Trip Serving Zones, the real-time infrastructure behind each ride or delivery that processes location data and launches millions of predictions within milliseconds. Simultaneously, the company has started testing Trainium3 for training some models that assist in driver or courier assignment, estimating arrival times, and personalizing delivery recommendations. In other words, there’s a distinction between the immediate operation of the service and the learning layer that improves these decisions over time.

Graviton4 Gains Ground in a Critical Part of the Business

The most mature aspect of the announcement pertains to Graviton4. AWS explains that Uber is already shifting more of its Trip Serving Zones workloads to this architecture to reduce energy consumption, enable rapid scaling during demand peaks, and lower latency in operations where just a few milliseconds can directly impact user experience. Uber summarizes this with a revealing phrase: “milliseconds matter.” In a business where the platform must match supply and demand in real time, computing efficiency is no small detail.

This aligns with the logic of Graviton within AWS. The Graviton family, based on Arm architecture, has been positioned for years as an option offering a better performance-to-efficiency ratio for certain cloud workloads. For Uber, this doesn’t mean a complete overhaul of its infrastructure but rather a targeted expansion over one of its most sensitive global systems. This offers an important hint: Uber is not using AWS solely as a generic capacity platform but as a hardware platform that offers specific advantages for finely tuned production workloads.

Trainium3 Enters Pilot Phase and Reinforces AI Strategy

The other part of the deal is more experimental but likely more strategic in the medium term. Uber has begun pilot testing the training of some AI models on Trainium3, AWS’s accelerator for training. AWS states that these models analyze data from billions of trips and deliveries to optimize matching between users and drivers, estimate arrival times, and personalize recommendations. For now, Amazon’s communication makes it clear this is a pilot—not a full migration of Uber’s entire AI training layer to Trainium3.

This nuance is quite important. In the current AI chip market, many companies announce tests, pilots, or validations with alternative accelerators, but few immediately replace their main workflows entirely. In Uber’s case, the message seems more cautious: Trainium3 is being positioned as an option to build a technological foundation for faster and more efficient predictions and models, but without framing it as a break from existing AI infrastructures.

An AWS Partnership That Doesn’t Abandon Multi-Cloud Strategy

The announcement is also noteworthy because it doesn’t contradict Uber’s multi-cloud strategy formalized in 2023. That year, Uber signed seven-year agreements with Oracle Cloud Infrastructure and Google Cloud as part of its gradual transition away from proprietary data centers. Oracle announced a strategic seven-year partnership to accelerate Uber’s cloud migration, and Google concurrently announced an expansion of its relationship with Uber to move applications and data from on-premise data centers to Google Cloud.

The key, therefore, is that Uber isn’t abandoning its multi-cloud approach but refining it. Google and Oracle remain part of its long-term architecture, while AWS gains strength in a domain where it can make a real difference today: combining general-purpose compute optimized with Graviton4 and a potential AI training route with Trainium3. Rather than an exclusive commitment to a single provider, what’s evident is a more selective allocation of workloads based on cost, latency, elasticity, and hardware availability.

This also aligns with Uber’s recent trajectory. The company has explained in technical publications that its infrastructure modernization aims for increased productivity, faster engineering cycles, and better cost efficiency. In that context, expanding AWS for a critical component of its business isn’t a script change but a logical continuation of a distributed architecture where each cloud can contribute something unique.

AWS Gains a Valuable Reference in Proprietary Chips

For AWS, this partnership carries symbolic and commercial value. Uber is not a small startup or trial project: it’s a global, real-time application with millions of daily users and strict demands for availability and latency. Having such a company push more operational load onto Graviton4 and begin testing Trainium3 provides Amazon with a powerful reference for its proprietary silicon strategy. In essence, it’s not just about selling cloud capacity but demonstrating that its chips can support critical, large-scale applications.

Ultimately, this announcement highlights an increasingly visible trend in cloud infrastructure: competition is no longer only among providers but also between platforms with differentiated silicon. Uber needs speed, elasticity, efficiency, and a solid foundation to enhance its models further. AWS is offering Graviton4 for operational workloads and Trainium3 for experimental purposes. While it’s not yet a total overhaul, it’s a clear step toward a future where the chip underneath can influence the competitive edge.

FAQs

What exactly will Uber use from AWS under this new agreement?
Uber will expand its use of AWS Graviton4 instances for more of its Trip Serving Zones workloads and has begun pilot testing AI models on AWS Trainium3.

What are Uber’s Trip Serving Zones?
They are part of the real-time infrastructure that manages each trip or delivery, processing location data and generating millions of predictions within milliseconds to match users with drivers or couriers.

Does this mean Uber is abandoning Google Cloud or Oracle Cloud?
No. Uber maintains a multi-cloud strategy. In 2023, it signed seven-year agreements with Oracle and Google Cloud to migrate away from its own data centers, and this AWS news is more about a selective workload expansion rather than replacing those partnerships.

Will Trainium3 become Uber’s main AI platform?
Not for now. AWS and Uber describe it as a pilot for certain AI models. It’s an important move but still in testing, not a full migration of Uber’s AI workloads.

Scroll to Top