Learn about the plans of market leaders in streaming VR and AR content from the cloud in this panel discussion. From enterprise use cases to streaming VR to the 5G edge, panelists will describe the state-of-the-art and challenges to making XR truly mobile.
We'll examine the challenges for telecommunications companies of harvesting the considerable computational capacity of modern GPU architectures. One issue is that low latency inference requires small batch sizes, which are inherently detrimental to Tensor Core performance. Another involves efficient coefficient reuse, which demands very large matrix-matrix multiplications, while feedforward DNNs typically used for telecommunications ML have relatively small vector-matrix multiplications. We'll discuss our approach, which aims to provide low latency with significantly higher performance by improving use of computation capacity available in Tensor Cores.