Home / Technology / Nvidia's STX: AI Storage Revolution
Nvidia's STX: AI Storage Revolution
17 Mar
Summary
- Nvidia STX architecture targets AI context memory bottlenecks.
- STX offers 5x token throughput and 4x energy efficiency.
- Partnership ecosystem to build AI-native infrastructure.

Nvidia's GTC 2026 introduction of the BlueField-4 STX reference architecture addresses a critical storage bottleneck in AI agent performance. This modular design inserts a dedicated context memory layer between GPUs and traditional storage, targeting the inefficiency of key-value cache data access during inference. STX claims a significant 5x increase in token throughput, 4x improvement in energy efficiency, and 2x faster data ingestion compared to conventional CPU-based storage.
This architecture is not a standalone Nvidia product but a blueprint for its storage partner ecosystem, enabling them to build AI-native infrastructure. The system leverages a new storage-optimized BlueField-4 processor, integrating Nvidia's Vera CPU and ConnectX-9 SuperNIC, running on Spectrum-X Ethernet networking and programmable via the DOCA software platform. Nvidia is also expanding DOCA with a new component, DOCA Memo, to further optimize storage for agentic AI workloads.
Major storage vendors like Dell, HPE, IBM, and NetApp, alongside AI-native cloud providers such as CoreWeave and Oracle Cloud Infrastructure, are collaborating on STX-based infrastructure. These platforms are anticipated to be available in the second half of 2026, signaling a shift towards treating the storage layer as a crucial element in enterprise AI planning.
IBM exemplifies this trend, integrating its Storage Scale System 6000 into Nvidia's DGX platforms and announcing an expanded collaboration with Nvidia on GPU-accelerated data processing. A proof of concept with Nestlé demonstrated substantial improvements in data refresh cycles, highlighting the performance gains achievable by optimizing the data layer for AI workloads.




