Red Hat has developed the Red Hat AI Factory platform with NVIDIA with NVIDIAavailable now, which combines Red Hat AI Enterprise and NVIDIA AI Enterprise to deliver an end-to-end AI solution optimized for organizations deploying AI at scale.
This platform enables IT operations teams to streamline traditional infrastructure management while addressing the changing demands of the AI stack. Red Hat AI Factory with NVIDIA advances production AI, delivering a software platform for AI factories, running on an accelerated computing infrastructure that drives higher performance for NVIDIA models and GPUs that power the inference stack.
The platform supports AI factory infrastructure from leading system manufacturers, including Cisco, Dell Technologies, Lenovo, and Supermicro. This way, IT administrators and operations teams can scale and maintain AI deployments with operational rigor and the same predictability as any other enterprise workload.
Red Hat AI Factory with NVIDIA integrates the open source collaboration, engineering and support expertise of Red Hat and NVIDIA to provide a scalable foundation for AI deployments in any environment, whether on-premises, the cloud or the edge. It includes essential capabilities for high-performance AI inference, model tuning, personalization, and agent deployment and management, with a focus on security.
This translates, by maintaining architectural control, into a reduction in time to value, as the solution facilitates the advancement of production AI with optimized workflows and instant access to pre-configured models. Among them, those from the IBM Granite family, NVIDIA Nemotron and the NVIDIA Cosmos open models. Additionally, it enables greater alignment of models with enterprise data using NVIDIA NeMo.
It also optimizes infrastructure usage and boosts inference performance with a unified, high-performance service stack. Red Hat AI Factory with NVIDIA delivers built-in observability capabilities, as well as leveraging Red Hat AI inference capabilities powered by vLLM, NVIDIA TensorRT-LLM, and NVIDIA Dynamo to meet AI service level objectives. This way, companies can reduce the total cost of ownership for AI.
Platform-enabled Intelligent GPU Orchestration, on the other hand, enables on-demand access to GPU resources through intelligent orchestration and clustered infrastructure, with automatic checkpointing to protect long-running jobs, and helps maintain more predictable compute costs in dynamic environments.
Chris Wright, CTO and Vice President of Global Engineering at Red Hathas highlighted that «Moving from AI experimentation to enterprise-wide, industrial-scale production requires a fundamental shift in how we manage the AI computing stack. We are accelerating the path to deploy AI and quickly move to production using Red Hat AI Factory with NVIDIA. With a stable, high-performance foundation driven by our renowned hybrid cloud proposition, we are enabling our customers to take charge of their AI strategy and scale with the same rigor they apply to their core IT platforms.«.
Justin Boitano, Vice President of Enterprise AI Platforms at NVIDIAhas commented that «Enterprises are building AI factories that turn data into intelligence at scale during inference, requiring production-grade infrastructure and software that spans hybrid cloud. Red Hat AI Factory with NVIDIA provides the software foundation that helps organizations keep pace with rapid infrastructure innovation while reliably building and deploying the next generation of agentic AI applications.«.
