Red Hat Expands Collaboration With Aws To Power Enterprise-grade Generative Ai

37 Min(s) Ago    👁 50
 

Red Hat, a global leader in open source solutions, has announced an expanded collaboration with Amazon Web Services AWS to deliver enterprise-grade generative AI gen AI capabilities on AWS using Red Hat AI and AWS AI silicon. The partnership is designed to give IT decision-makers the flexibility to run high-performance, efficient AI inference at scale, independent of underlying hardware.

The growing demand for gen AI and scalable inference is prompting organizations to rethink their IT infrastructure. According to IDC, by 2027, 40 of organizations are expected to use custom silicon, including ARM processors or AI/ML-specific chips, to meet the increasing need for performance optimization, cost efficiency, and specialized computing. This highlights the importance of solutions that improve processing power, reduce costs, and accelerate innovation for high-performance AI applications.

The collaboration brings together Red Hat's comprehensive AI platform capabilities with AWS cloud infrastructure and AI chipsets, including AWS Inferentia2 and AWS Trainium3. Red Hat AI Inference Server, powered by vLLM, will be optimized to run on AWS AI chips, providing a common inference layer that supports any gen AI model. This enables higher performance, lower latency, and cost-effective scaling of production AI deployments, with up to 30-40 better price performance than comparable GPU-based Amazon EC2 instances.

Disclaimer: We are a news aggregator. See full disclaimer here.