GTC ON-DEMAND

 
SEARCH SESSIONS
SEARCH SESSIONS

Search All
 
Refine Results:
 
Year(s)

SOCIAL MEDIA

EMAIL SUBSCRIPTION

 
 

GTC ON-DEMAND

Presentation
Media
Abstract:
Learn our solutions for increasing GPU resource utilization on an on-premise DGX-2 node and public clouds. In this talk we present our operational experiences of a set of multi-tenant deep learning workloads selected through an open competition. To host them we use and extend the Backend.AI framework as the resource and computation manager. While tailored for both educational and research-oriented workloads, it offers a topology-aware multi-GPU resource scheduler combined with fractional GPU scaling implemented via API-level CUDA virtualization, achieving higher GPU utilization compared to vanilla setups.
Learn our solutions for increasing GPU resource utilization on an on-premise DGX-2 node and public clouds. In this talk we present our operational experiences of a set of multi-tenant deep learning workloads selected through an open competition. To host them we use and extend the Backend.AI framework as the resource and computation manager. While tailored for both educational and research-oriented workloads, it offers a topology-aware multi-GPU resource scheduler combined with fractional GPU scaling implemented via API-level CUDA virtualization, achieving higher GPU utilization compared to vanilla setups.  Back
 
Topics:
Data Center & Cloud Infrastructure, GPU Virtualization, Deep Learning & AI Frameworks
Type:
Talk
Event:
GTC Silicon Valley
Year:
2019
Session ID:
S9406
Download:
Share:
 
Abstract:
Learn our solutions and experiences for scaling up and increasing utilization of GPU resources on DGX and cloud platforms. Our solution, Backend.AI, is the GPU-first framework to host machine learning workloads and fully integrates with NGC. We will describe how it supports diverse-sized computations using both aggregation of multiple GPUs and fractional scaling of GPUs, and how it reduces modelling time for data scientists through a public competition using the DGX platform.
Learn our solutions and experiences for scaling up and increasing utilization of GPU resources on DGX and cloud platforms. Our solution, Backend.AI, is the GPU-first framework to host machine learning workloads and fully integrates with NGC. We will describe how it supports diverse-sized computations using both aggregation of multiple GPUs and fractional scaling of GPUs, and how it reduces modelling time for data scientists through a public competition using the DGX platform.  Back
 
Topics:
Artificial Intelligence and Deep Learning
Type:
Talk
Event:
AI Conference Korea
Year:
2019
Session ID:
SKR9115
Download:
Share:
 
 
Previous
  • Amazon Web Services
  • IBM
  • Cisco
  • Dell EMC
  • Hewlett Packard Enterprise
  • Inspur
  • Lenovo
  • SenseTime
  • Supermicro Computers
  • Synnex
  • Autodesk
  • HP
  • Linear Technology
  • MSI Computer Corp.
  • OPTIS
  • PNY
  • SK Hynix
  • vmware
  • Abaco Systems
  • Acceleware Ltd.
  • ASUSTeK COMPUTER INC
  • Cray Inc.
  • Exxact Corporation
  • Flanders - Belgium
  • Google Cloud
  • HTC VIVE
  • Liqid
  • MapD
  • Penguin Computing
  • SAP
  • Sugon
  • Twitter
Next