A Blueprint for LLM Cluster Architecture: Scaling to the World’s Largest Deployments
- Location
- San Jose, CA
- Date
- –
- Speakers:
- Mehnaz Mahbub
- Alok Srivastav
In this session, we will be revealing the blueprints of modern data center infrastructure supporting the training of Large Language Models. First, we will provide an overview of designing cluster-level architecture. Then, we will showcase an onsite deployment of one of the largest AI training clusters in the world.
Transform the Retail Experience: Architecting LLM Inferencing Systems for Edge Deployment
- Location
- San Jose, CA
- Date
- –
- Speakers:
- Jeff Sharpe
As Large Language Models are rapidly integrated into every industry, Edge AI is crucial for enabling faster, more secure, and efficient data processing, particularly for applications where real-time analytics and decision-making are essential. In this session, we explore the transformative potential of deploying powerful systems at the Edge to enhance in-store experiences.
Accelerating AI Data Pipelines: A Case Study Featuring 25 Petabytes of Tiered Storage
- Location
- Virtual
- Speakers:
- William Li
- Paolo Basilio
Storage optimized for AI workloads must have high performance throughput to stage data on GPU cluster servers while also having a very large, cost-effective, capacity optimized mass storage tier to collect, process and label large data sets needed for the AI model training. In this session, Supermicro will discuss AI storage solutions using high performance flash-based storage servers and high-capacity disk storage servers with file and object storage solutions from partners such as Weka.io, OSNexus, Quantum ActiveScale and Qumulo.
Activating AI Infrastructure Breakthroughs in Density and Efficiency With Total Liquid-Cooling Solutions
- Location
- Virtual
- Speakers:
- Nathan Mallamace
- Jae Lee
In this session, we explore why many modern data centers are starting to opt for liquid-cooling infrastructure. We will briefly cover liquid-cooling key components and provide an overview on liquid-cooling deployments. We highlight a few examples of AI power costs, ranging from real-time inferencing workloads to training LLM foundational models. Then, we will present a cost-benefit analysis of liquid versus air-cooled AI infrastructure.
Mehnaz Mahbub
Sr. Solutions Manager
Alok Srivastav
Sr. Solutions Manager
Jeff Sharpe
Sr. Director, Edge AI Solutions
William Li
Director, Solution Management
Paolo Basilio
Senior Storage Architect
Nathan Mallamace
Strategic Business Development Manager
Jae Lee
Product Manager, Thermal Solutions
Supermicro at GTC 2024
Supermicro was back in-person at GTC 2024, the #1 AI conference for developers and industry professionals. Our booth featured a portfolio of cutting-edge AI and accelerated computing solutions, including our latest GPU systems and data center racks. Attendees caught a glimpse of the future of AI hardware with a showcase of next generation technologies. Supermicro experts spoke about the key technologies that define the AI data center, such as liquid cooling and cluster-level architecture for LLM training.
The event has concluded. Hope to see you next year!