SEARCH SESSIONS

Search All
 
Refine Results:
 
Year(s)

SOCIAL MEDIA

EMAIL SUBSCRIPTION

 
 

GTC ON-DEMAND

AI Application Deployment and Inference
Presentation
Media
Accelerate TensorFlow Inference with New TensorRT Integration
Abstract:
TensorFlow is an open source software library for numerical computation using data flow graphs. NVIDIA TensorRT is an inference optimizer and runtime for runtime deployment. TensorRT provides optimizations for deep neural networks and uses reduced precision to increase throughput, reduce latency, while maintaining accuracy. Today we announced tighter integration in TensorFlow for TensorRT through with new TensorFlow APIs, sub-graph optimizations and INT8 calibration to automatically leverage Tensor Cores on Volta GPUs. TensorRT delivers 2.5x faster inference throughput compared to inference without TensorRT. In this session, NVIDIA developers will use an example based workflow to show how to use this new capability.
 
Topics:
AI Application Deployment and Inference, Deep Learning and AI Frameworks
Type:
Talk
Event:
GTC Silicon Valley
Year:
2018
Session ID:
S81009
Streaming:
Download:
Share: