VAST Data and NVIDIA integrate the “AI OS” into the server: CNode-X aims to simplify the AI stack and accelerate RAG and vector search

The Artificial Intelligence infrastructure is entering a new phase: the battle is no longer just about GPU power but now focuses on the data pathway. In this context, VAST Data and NVIDIA have announced an integration that directly addresses the most common pain point in enterprise deployments: the complexity of connecting storage, databases, analytics, and AI services as if they were separate pieces. The partnership materializes in VAST CNode-X, a new offering to deliver the VAST AI Operating System (AI OS) directly to NVIDIA GPU servers and transform it into a ‘first-class’ layer within accelerated clusters.

The idea is as simple as it is ambitious: if training, inference, and RAG-like systems depend on moving data back and forth, then real performance is determined at the data path. VAST claims that, with AI OS running on NVIDIA-powered servers, organizations can eliminate bottlenecks and run ingestion, retrieval, analytics, and inference as a unified, coherent system. Instead of piecing together separate tools—storage on one side, databases on another, and a separate AI stack—this proposal aims to concentrate those capabilities into a unified, production-oriented stack.

CNode-X: when the server stops being “just compute”

VAST describes CNode-X as a new generation of NVIDIA-Certified Systems, designed to transform how AI infrastructure is built and operated. The approach goes beyond simply “connecting” fast storage to a GPU cluster: the innovation is that AI OS is available directly on the hardware, providing high-performance storage services to accelerated clusters and integrating the data operating system within the environment itself.

The declared goal is to optimize a set of use cases that are now central to any serious AI operation: AI pipelines, high-performance analytics, vector search, RAG, and agent-based workloads. In practice, this means reducing the ‘gluing’ work that often appears when transitioning from a pilot project to a durable system: different configurations, separate consoles, duplicated identities, and multiple teams trying to make the stack behave as a single product.

Renen Hallak, founder and CEO of VAST Data, frames this evolution as the realization of a goal pursued for a decade: a system capable of “refining data into intelligence and action” continuously. His argument is that accelerating compute and data pathways within AI OS offers a more direct route to bringing recovery, analytics, and agent workflows into production, avoiding the project dying in the integration phase.

Jensen Huang and “persistent memory” for agents working over days

NVIDIA, meanwhile, positions the initiative within its strategy to “reinvent” the pillars of compute for AI. Jensen Huang, founder and CEO, highlights that CNode-X will be accelerated by CUDA “at every layer” to give AI agents persistent memory, enabling them to tackle complex problems over days or weeks and, eventually, over years, without “forgetting” the context. This is a key message because it points to an industry trend: agents require not only powerful models but also a data system that maintains state, history, and large-scale information retrieval.

Additionally, Cisco joins the story as one of the commercialization partners. Jeetu Patel, President and CPO of Cisco, suggests that companies need a reliable path to operationalize AI at scale. His view is that collaboration with VAST and NVIDIA will bring to market an end-to-end accelerated platform on Cisco infrastructure, helping move from experimentation to “always-on” systems more quickly and with greater confidence.

OEMs: availability through Cisco, HPE, and Supermicro

VAST indicates that CNode-X servers will be available via manufacturers such as Cisco and Supermicro, with the sector’s coverage adding that they will also come from HPE and other OEMs. This detail is significant: in infrastructure, adoption accelerates when the product is purchased “off the shelf” through conventional channels, with certified support and configurations.

A fitting context: “AI networking” no longer resembles Internet networking

The news is accompanied by a fact that helps explain why storage is becoming a central focus again. A recent Backblaze analysis of neocloud traffic suggests that the network behavior of AI workloads differs from traditional cloud: “less many-to-many and more sustained high-bandwidth relationships between specialized systems.” In this scenario, Backblaze suggests that storage, compute, and network will tend to align even more tightly, driven by the need to keep GPUs busy and pipelines flowing.

This perspective aligns with CNode-X’s commercial argument: when the bottleneck is the data, the value isn’t just in adding more GPUs but in reducing operational latency, simplifying the stack, and preventing infrastructure from becoming an endless integration project.

Polaris: the layer for handling distributed AI as “a single fleet”

The expansion of the VAST-NVIDIA alliance was announced alongside the launch of Polaris, a new multi-cloud and hybrid offering from VAST. The goal of Polaris is to turn distributed deployments into a single, fleet-scale platform for managing AI infrastructure and data wherever training and inference workloads run. The message aligns with market trends: AI doesn’t reside in one data center or cloud, and increasingly, organizations want a unified control plane without multiplying tools.

Altogether, CNode-X and Polaris reflect a shared ambition: to turn AI infrastructure from a collection of disparate parts into a data and execution operating system, ready for production.


Frequently Asked Questions (FAQ)

What is VAST CNode-X and what role does it play in a NVIDIA GPU cluster?
VAST CNode-X is a GPU-powered server proposition that integrates VAST AI OS directly into the hardware, providing high-performance data services and simplifying the operation of AI pipelines, RAG, and vector search within accelerated clusters.

Why does VAST emphasize “unifying” storage, databases, and the AI stack?
Because in many enterprise deployments, the biggest obstacle isn’t the model itself but the integration: separate tools, multiple consoles, and data flows with latency. Unifying the stack aims to reduce complexity and speed up moving from pilot to production.

What does “persistent memory for AI agents” mean in this context?
It refers to maintaining the state and context of agents over extended periods, supported by an optimized data and recovery layer, allowing agents to perform complex tasks without losing relevant information.

What does Polaris contribute to a hybrid multicloud AI strategy?
Polaris aims to be a control plane that manages distributed infrastructure and data—spanning data centers and clouds—as a single “fleet,” making operations coherent when training and inference happen across different environments.

Scroll to Top