Summary

Pure Storage has achieved storage partner validation for NVIDIA OVX servers. The addition of OVX reference architectures to AIRI and FlashStack helps make Pure Storage the go-to provider for enterprise AI-ready infrastructure solutions.

image_pdfimage_print

At NVIDIA GTC 2024, Pure Storage announced its close collaboration with NVIDIA to achieve storage partner validation for NVIDIA OVX servers powered by NVIDIA L40S GPUs and FlashBlade//S™. This new reference architecture validation provides enterprises with greater GPU server choice and more immediate availability of proven AI infrastructure for fast and efficient small model training, fine-tuning, and inference workloads.

The addition of Pure Storage’s validated NVIDIA OVX reference architectures to our recent AIRI® built on NVIDIA DGX BasePOD and new FlashStack® for AI solutions helps make Pure Storage the go-to provider for enterprise AI-ready infrastructure solutions.

What is NVIDIA OVX Validation?

NVIDIA OVX is an advanced computing environment engineered to run Omniverse applications, however the certification can be applied to other infrastructure to indicate its performance and compatibility with AI applications. The OVX certification indicates that AI infrastructure, such as data storage, is optimized for demanding AI workloads, meeting the rigorous standards required by OVX servers, including reliability, performance, and compatibility. Through close collaboration, we designed these solutions specifically to integrate with OVX servers powered by NVIDIA L40s GPUs, for an infrastructure choice enterprises can count on for demanding AI workloads such as generative AI, real-time simulations, and 3D.

Complementing Validated Enterprise Infrastructure for GenAI RAG

Pure Storage launched the first and original AIRI in 2018 and we’ve since provided customers with steady innovations with the introduction of FlashBlade//S (the storage foundation for AIRI), as well as Cisco Validated Designs with FlashStack for AI.

Our goal: to give customers the best options for full-stack, ready-to-run AI infrastructure.

The recent demand for new AI workloads like Generative AI and retrieval-augmented generation (RAG) to customize large language models (LLMs) for specific domain or company needs has elevated the demand for efficient, powerful AI compute, storage, and networking solutions.

These new AI use cases require high-performance and highly efficient all-flash storage to maximize GPU utilization and AI productivity. Pure Storage’s data storage platform for AI ensures that multimodal performance addresses the common payload in an AI data pipeline with performance and capacity-optimized DirectFlash® technology. Our products require 80% less energy than alternatives and that shifts availability of power for more GPUs per rack.

With the Pure Storage Platform for AI, customers are able to:

  1. Accelerate model training and inferencing
  2. Maximize operational efficiency
  3. Deliver cost and energy efficiency at scale
  4. Achieve ultimate reliability and future-proof AI storage

MLOps and Vertical Full-stack Solutions

The difficulty in getting AI initiatives off the ground spans multiple domains. Fundamental challenges can be due to:

  • Legacy compute systems
  • Lack of skilled people
  • Limited budgets
  • Silos of legacy data storage that don’t keep powerful AI-optimized GPUs fully productive

Domain expertise beyond infrastructure and storage is needed around the use and deployment of MLOps applications from providers such as Red Hat OpenShift, Weights & Biases, Run:ai, Ray, and Anyscale, which integrate with NVIDIA AI Enterprise software, including NVIDIA NeMo and new NVIDIA NIM and NeMo Retriever microservices. Some industries require vertical-specific applications, including financial services applications like KDB.AI, Raft, and Milvus. The healthcare and life sciences stack can include applications from NVIDIA Clara like MONAI for medical imaging.

New validated full-stack solutions, including MLOps and vertical stacks from Pure Storage, help fast-track IT and AI infrastructure teams with solutions that have gone through thorough testing, documentation, and integration so that AI development and data science teams can be off and running faster and with less risk.

Future-proof Storage for Uncertain AI Growth

The Pure Storage platform for AI is future-proof. Companies can invest in AI infrastructure that can grow over time without fear of storage that is out of date and can’t keep up with fast-growing AI and data science team needs. Pure Storage also provides a subscription cloud model for consumption called Evergreen//One™ that makes consumption of storage for AI more cloud-like but with the advantages of on-premises, including continuous innovation, financial flexibility, and operational agility.

Learn more about how Pure Storage can help you accelerate adoption of AI as well as the benefits that AI can bring to your business.