Red Hat, a global leader in open source solutions, has launched Red Hat AI Factory with NVIDIA, a jointly developed software platform that integrates Red Hat AI Enterprise and NVIDIA AI Enterprise. This solution offers a comprehensive AI approach, optimized for organizations deploying artificial intelligence at scale. The introduction of Red Hat AI Factory with NVIDIA reinforces the collaboration between both companies, accelerating the availability of AI innovations for enterprise customers starting today, with support from day one for NVIDIA hardware architectures.
It is expected that enterprise AI spending will surpass one trillion dollars by 2029, driven largely by agentic AI applications. In this context, companies seek solutions that manage high-density agentic workflows and address the growing demands for AI infrastructure and inference. The new platform allows IT teams to optimize the management of both traditional infrastructure and the dynamic requirements of AI ecosystems.
Red Hat AI Factory with NVIDIA accelerates production AI adoption by offering a comprehensive platform for AI factories, built on accelerated computing infrastructure that enhances model performance and NVIDIA GPUs responsible for inference processing. The platform supports AI factory infrastructure from leading vendors such as Cisco, Dell Technologies, Lenovo, and Supermicro, making it easier for IT leaders and operations teams to scale and maintain AI deployments with the same predictability and rigor as any other enterprise workload.
This co-designed software platform combines open-source collaboration, engineering, and support expertise from both Red Hat and NVIDIA to deliver a reliable enterprise solution. Red Hat AI Factory with NVIDIA provides a highly scalable foundation for AI deployments across any environment—on-premises, cloud, or edge. It includes essential capabilities for high-performance AI inference, model tuning, personalization, deployment, and agent management, with a focus on security. This allows organizations to maintain architectural control from data centers to public clouds, leading to:
- Reduced time to value: This solution streamlines production AI workflows with optimized processes and instant access to preconfigured models, including the IBM Granite family with legal support, NVIDIA Nemotron, and open models like NVIDIA Cosmos, delivered as NVIDIA NIM microservices. It also enables organizations to better align models with business data using NVIDIA NeMo, reducing tuning time and costs.
- Optimized performance and cost: It maximizes infrastructure use and enhances inference performance through a unified, high-performance service stack. Red Hat AI Factory with NVIDIA offers integrated observability features and leverages inference capabilities powered by vLLM, NVIDIA TensorRT-LLM, and NVIDIA Dynamo to meet strict AI service level objectives. This helps organizations lower the total cost of ownership (TCO) for AI by optimizing the connection between models and NVIDIA GPUs.
- Smart GPU orchestration: It provides on-demand access to GPU resources through intelligent orchestration and pooled infrastructure, with automatic checkpoints to protect long-running jobs and keep compute costs predictable in dynamic environments.
- Enhanced enterprise security: By leveraging the flexible and robust foundation of Red Hat Enterprise Linux, organizations benefit from advanced security and compliance capabilities built-in from the outset, helping to reduce risk, save time, and mitigate downtime. This creates a reinforced security base for mission-critical AI workloads requiring isolation and continuous verification. NVIDIA DOCA microservices are built on this foundation, creating a zero-trust architecture and providing runtime AI security across the entire infrastructure.

