跳转到主要内容
Dates
Location
San Jose, CA and Virtual

Defining the future of AI Infrastructure

[YouTube] Supermicro AI Infrastructure GTC24 Update: AI Rack Architecture, Liquid-Cooling, AI Storage, Edge AI
Supermicro Sessions at GTC 2024

A Blueprint for LLM Cluster Architecture: Scaling to the World’s Largest Deployments

Location
San Jose, CA
Date
  • Speakers:
    • Mehnaz Mahbub
    • Alok Srivastav

In this session, we will be revealing the blueprints of modern data center infrastructure supporting the training of Large Language Models. First, we will provide an overview of designing cluster-level architecture. Then, we will showcase an onsite deployment of one of the largest AI training clusters in the world.

View on GTC Site

Transform the Retail Experience: Architecting LLM Inferencing Systems for Edge Deployment

Location
San Jose, CA
Date
  • Speakers:
    • Jeff Sharpe

As Large Language Models are rapidly integrated into every industry, Edge AI is crucial for enabling faster, more secure, and efficient data processing, particularly for applications where real-time analytics and decision-making are essential. In this session, we explore the transformative potential of deploying powerful systems at the Edge to enhance in-store experiences.

View on GTC Site

Accelerating AI Data Pipelines: A Case Study Featuring 25 Petabytes of Tiered Storage

Location
Virtual
  • Speakers:
    • William Li
    • Paolo Basilio

Storage optimized for AI workloads must have high performance throughput to stage data on GPU cluster servers while also having a very large, cost-effective, capacity optimized mass storage tier to collect, process and label large data sets needed for the AI model training. In this session, Supermicro will discuss AI storage solutions using high performance flash-based storage servers and high-capacity disk storage servers with file and object storage solutions from partners such as Weka.io, OSNexus, Quantum ActiveScale and Qumulo.

View on GTC Site

Activating AI Infrastructure Breakthroughs in Density and Efficiency With Total Liquid-Cooling Solutions

Location
Virtual
  • Speakers:
    • Nathan Mallamace
    • Jae Lee

In this session, we explore why many modern data centers are starting to opt for liquid-cooling infrastructure. We will briefly cover liquid-cooling key components and provide an overview on liquid-cooling deployments. We highlight a few examples of AI power costs, ranging from real-time inferencing workloads to training LLM foundational models. Then, we will present a cost-benefit analysis of liquid versus air-cooled AI infrastructure.

View on GTC Site

Mehnaz Mahbub, Sr. Solutions Manager at Supermicro

Mehnaz Mahbub

Sr. Solutions Manager

Alok Srivastav, Sr. Solutions Manager at Supermicro

Alok Srivastav

Sr. Solutions Manager

Jeff Sharpe, Sr. Director of Edge AI Solutions at Supermicro

Jeff Sharpe

Sr. Director, Edge AI Solutions

William Li, Director of Solution Management at Supermicro

William Li

Director, Solution Management

Paolo Basilio, Senior Storage Architect at Supermicro

Paolo Basilio

Senior Storage Architect

Nathan Mallamace, Strategic Business Development Manager at Supermicro

Nathan Mallamace

Strategic Business Development Manager

Jae Lee, Product Manager, Thermal Solutions at Supermicro

Jae Lee

Product Manager, Thermal Solutions

Supermicro at GTC 2024

Supermicro was back in-person at GTC 2024, the #1 AI conference for developers and industry professionals. Our booth featured a portfolio of cutting-edge AI and accelerated computing solutions, including our latest GPU systems and data center racks. Attendees caught a glimpse of the future of AI hardware with a showcase of next generation technologies. Supermicro experts spoke about the key technologies that define the AI data center, such as liquid cooling and cluster-level architecture for LLM training.

The event has concluded. Hope to see you next year!

Learn More

Rendering of Supermicro’s Booth at GTC 2024