Solidigm focused on the evolving role of storage in AI, specifically highlighting its significance in the AI data pipeline through 2025. The presentation emphasizes that the AI workflow involves a series of distinct tasks, each with unique demands on hardware, and that it's often distributed among different organizations. The presentation identifies six primary stages in this pipeline: ingestion, data preparation, model training, fine-tuning, inference, and archiving. Solidigm's perspective underscores the increasing data intensity of these steps, especially during data ingestion, inference, and archiving.
Solidigm explores how storage choices impact these stages, distinguishing between direct-attached storage within GPU servers and network-attached storage. Direct-attached storage is optimized for performance-intensive tasks, while network-attached storage is used for larger datasets and offers capacity and cost-effectiveness. Ace Stryker highlights the rising importance of network storage, driven by advances in network bandwidth and the growing size and complexity of AI models and larger datasets for RAG. The key takeaway is that great storage facilitates larger models, longer interactions, and improved outputs on the inference side.
Finally, the presentation showcases a collaboration with Metrum AI, presenting a real-world demo that addresses the challenges of data-intensive inference by offloading model weights and RAG data to SSDs. This allows running 70 billion parameter models on less powerful hardware, which would have been impossible without storage-based offloading, saving on costs and hardware requirements. The demo emphasizes the potential of SSDs to enhance performance in AI applications by reducing GPU memory usage. The collaboration offers insights into the benefits of leveraging storage in AI, as the partnership also showed that offloading to SSD could provide the same or better performance as DRAM in this application of retrieval augmented generation.
Presented by Ace Stryker, Director of Market Development, Solidigm. Recorded live in Santa Clara, California, on April 23, 2025, as part of AI Infrastructure Field Day. Watch the entire presentation at https://techfieldday.com/appearance/solidigm-presents-at-ai-infrastructure-field-day-2/or https://techfieldday.com/event/aiifd2/ for more information.
Up Next in AI Infrastructure Field Day 2
-
Industry First SSDs Tackle AI Challen...
Solidigm's Director of Leadership Narrative and Evangelist, Scott Shadley, introduces their new SSDs designed to tackle the challenges of AI infrastructure. Shadley begins by highlighting the industry-first status of their 122TB SSDs, emphasizing their potential to revolutionize storage solutions...
-
Affordable on premises LLM training a...
Phison's show their aiDAPTIV+ technology, designed to make on-premises AI processing more affordable and accessible, particularly for small to medium-sized businesses, governments, and universities. The core of their innovation lies in leveraging flash storage to offload the memory demands of lar...
-
GPU Memory Offload for LLM fine-tunin...
With aiDAPTIV+, Phison makes on-premises AI processing more accessible and affordable, especially for small and medium-sized businesses, government entities, and educational institutions. CTO Sebastien Jean explained that the primary goal of Phison’s solution is to facilitate fine-tuning of large...