Red Hat, a leading provider of open source solutions, announced an expanded collaboration with Amazon Web Services (AWS) to power enterprise-grade generative AI (gen AI) on AWS with Red Hat AI and AWS AI silicon.
With this collaboration, Red Hat focuses on empowering IT decision-makers with the flexibility to run high-performance, efficient AI inference at scale, regardless of the underlying hardware, according to the companies.
The rise of gen AI and subsequent need for scalable inference is pushing organizations to reevaluate their IT infrastructure.
Red Hat’s collaboration with AWS empowers organizations with a full-stack gen AI strategy by bringing together Red Hat’s comprehensive platform capabilities with AWS cloud infrastructure and AI chipsets, AWS Inferentia2 and AWS Trainium3.
Key aspects of the collaboration include:
- Red Hat AI Inference Server on AWS AI chips
- Enabling AI on Red Hat OpenShift
- Ease of access and deployment
- Upstream community contribution
Red Hat has a long history of collaboration with AWS to enable customers from the data center to the edge.
This latest milestone now aims to address the evolving needs of organizations as they integrate AI into their hybrid cloud strategies to achieve optimized, efficient gen AI outcomes, the companies said.
The AWS Neuron community operator is now available in the Red Hat OpenShift OperatorHub for customers using Red Hat OpenShift or Red Hat OpenShift Service on AWS.
Red Hat AI Inference Server support for AWS AI chips is expected to be available in developer preview in January 2026.
For more information about this news, visit www.redhat.com.