SEARCH SESSIONS

Search All
 
Refine Results:
 
Year(s)

SOCIAL MEDIA

EMAIL SUBSCRIPTION

 
 

GTC ON-DEMAND

AI Application Deployment and Inference
Presentation
Media
CNN Inference with cuDNN: Common Pitfalls and Best Practices
Abstract:
You may already use NVIDIA's cuDNN library to accelerate your deep neural network inference, but are you getting the most out of it to truly unleash the tremendous performance of NVIDIA's newest GPU architectures, Volta and Turing? We'll discuss how to avoid the most common pitfalls in porting your CPU-based inference to the GPU and demonstrate best practices in a step-by-step optimization of an example network. Learn how to deploy your deep neural network inference in both the fastest and most memory-efficient way, using cuDNN and Tensor Cores, NVIDIA's revolutionary technology that delivers groundbreaking performance in FP16, INT8 and INT4 inference on Volta and Turing.
 
Topics:
AI Application Deployment and Inference, Deep Learning and AI Frameworks
Type:
Talk
Event:
GTC Silicon Valley
Year:
2019
Session ID:
S9644
Streaming:
Download:
Share: