SEARCH SESSIONS

Search All
 
Refine Results:
 
Year(s)

SOCIAL MEDIA

EMAIL SUBSCRIPTION

 
 

GTC On-Demand

Deep Learning and AI
Presentation
Media
Opening Keynote
The 2018 GTC opening keynote is delivered by the NVIDIA Founder and CEO, Jensen Huang, speaking on the future of computing. ...Read More

The 2018 GTC opening keynote is delivered by the NVIDIA Founder and CEO, Jensen Huang, speaking on the future of computing.

  Back
 
Keywords:
Deep Learning and AI, GTC Silicon Valley 2018 - ID S8885
Streaming:
Share:
 
Opening Keynote
Don't miss this keynote from NVIDIA's Chief Scientist, Bill Dally, as he speaks on the future of computing.
Don't miss this keynote from NVIDIA's Chief Scientist, Bill Dally, as he speaks on the future of computing.  Back
 
Keywords:
Deep Learning and AI, GTC Israel 2018 - ID SIL8101
Streaming:
Share:
 
GTC Taiwan Opening Keynote 2018
The 2018 GTC Taiwan opening keynote is delivered by NVIDIA Founder and CEO, Jensen Huang.
The 2018 GTC Taiwan opening keynote is delivered by NVIDIA Founder and CEO, Jensen Huang.  Back
 
Keywords:
Deep Learning and AI, Autonomous Vehicles, AI and DL Research, GTC Taiwan 2018 - ID STW8000
Streaming:
Share:
 
Today's AI Industrial Revolution - What's Real, What's Next?
Many examples of AI are reported daily that enhance traditional products and servicesââ¬â but the benefits have only just begun to scratch the surface. Entire industries will be transformed and massive benefits will ...Read More

Many examples of AI are reported daily that enhance traditional products and servicesââ¬â but the benefits have only just begun to scratch the surface. Entire industries will be transformed and massive benefits will be realized in the next wave of AI deployment. Learn about the next generation of AI and how it will add over 60M jobs and $13 trillion to the global economy if policy makers, businesses and the AI community adopt the right strategies, initiatives and platforms to harness this incredible new technology.

  Back
 
Keywords:
Deep Learning and AI, GTC Washington D.C. 2018 - ID DC8212
Streaming:
Share:
Artificial Intelligence and Deep Learning
Presentation
Media
Opening Keynote
The GTC Europe 2018 opening keynote delivered by NVIDIA Founder and CEO, Jensen Huang, speaking on the future of computing. ...Read More

The GTC Europe 2018 opening keynote delivered by NVIDIA Founder and CEO, Jensen Huang, speaking on the future of computing.

  Back
 
Keywords:
Artificial Intelligence and Deep Learning, GTC Europe 2018 - ID E8001
Streaming:
Share:
Deep Learning and AI
Presentation
Media
GTC CHINA 2018 主题演讲
GTC CHINA 2018 主题演讲由 NVIDIA 创始人兼 CEO 黄仁勋(Jensen Huang)带来。
GTC CHINA 2018 主题演讲由 NVIDIA 创始人兼 CEO 黄仁勋(Jensen Huang)带来。  Back
 
Keywords:
Deep Learning and AI, AI for Business, AI for Gaming, RAPIDS, HPC and Supercomputing, GTC China 2018 - ID CH8100
Streaming:
Share:
 
Keynote
Don't miss this keynote from NVIDIA Founder & CEO, Jensen Huang, as he speaks on the future of computing. ...Read More

Don't miss this keynote from NVIDIA Founder & CEO, Jensen Huang, as he speaks on the future of computing.

  Back
 
Keywords:
Deep Learning and AI, Data Center and Cloud Infrastructure, Virtual Reality and Augmented Reality, Self-Driving Cars, Intelligent Video Analytics and Smart Cities, GTC Silicon Valley 2017 - ID S7820
Streaming:
Share:
 
Keynote
Don't miss this keynote from NVIDIA Founder & CEO, Jensen Huang, as he speaks on the future of computing.  ...Read More

Don't miss this keynote from NVIDIA Founder & CEO, Jensen Huang, as he speaks on the future of computing. 

  Back
 
Keywords:
Deep Learning and AI, Virtual Reality and Augmented Reality, Data Center and Cloud Infrastructure, Autonomous Vehicles, Intelligent Video Analytics and Smart Cities, GTC Israel 2017 - ID SIL7001
Streaming:
Share:
 
Keynote
 
Keywords:
Deep Learning and AI, GTC Taiwan 2017 - ID GTCT7101
Streaming:
Share:
Accelerated Analytics
Presentation
Media
Accelerated Computing: The Path Forward - Special Address
 
Keywords:
Accelerated Analytics, SIGGRAPH 2017 - ID SC1700
Share:
Leadership in AI
Presentation
Media
AI: Transforming Your Work and the World Now
Artificial intelligence is changing the world at an accelerating pace. AI has quickly jumped from research labs to business and consumer applications. In this keynote, Greg will share the latest developments in AI for transportation, robotics, manufa ...Read More
Artificial intelligence is changing the world at an accelerating pace. AI has quickly jumped from research labs to business and consumer applications. In this keynote, Greg will share the latest developments in AI for transportation, robotics, manufacturing, healthcare and government.  Back
 
Keywords:
Leadership in AI, GTC Washington D.C. 2017 - ID DC7114
Download:
Share:
AI in Healthcare
Presentation
Media
Harnessing AI in Healthcare
As computers outperform humans at complex cognitive tasks, disruptive innovation will increasingly remap the familiar with waves of creative destruction. And in healthcare, nowhere is this more apparent or imminent than at the crossroads of Radiology ...Read More
As computers outperform humans at complex cognitive tasks, disruptive innovation will increasingly remap the familiar with waves of creative destruction. And in healthcare, nowhere is this more apparent or imminent than at the crossroads of Radiology and the emerging field of Clinical Data Science. As leaders in our field, we must shepherd the innovations of cognitive computing by defining its role within diagnostic imaging, while first and foremost ensuring the continued safety of our patients. If we are dismissive, defensive or self-motivated - industry, payers and provider entities will innovate around us achieving different forms of disruption, optimized to serve their own needs. To maintain our leadership position, as we enter the era of machine learning, it is essential that we serve our patients by directly managing the use of clinical data science towards the improvement of carea position which will only strengthen our relevance in the care process.  Back
 
Keywords:
AI in Healthcare, GTC Washington D.C. 2017 - ID DC7240
Download:
Share:
Deep Learning and AI
Presentation
Media
Keynote
Don't miss this keynote from NVIDIA Founder & CEO, Jensen Huang, as he speaks on the future of computing.  ...Read More

Don't miss this keynote from NVIDIA Founder & CEO, Jensen Huang, as he speaks on the future of computing. 

  Back
 
Keywords:
Deep Learning and AI, Accelerated Analytics, AI Startup, GTC Japan 2017 - ID 1000
Streaming:
Share:
 
Keynote
Don't miss this keynote from NVIDIA Founder & CEO, Jensen Huang, as he speaks on the future of computing.  ...Read More

Don't miss this keynote from NVIDIA Founder & CEO, Jensen Huang, as he speaks on the future of computing. 

  Back
 
Keywords:
Deep Learning and AI, Autonomous Vehicles, Self-Driving Cars, Data Center and Cloud Infrastructure, GTC Europe 2017 - ID 23000
Streaming:
Share:
 
A New Computing Era
N/A ...Read More

N/A

  Back
 
Keywords:
Deep Learning and AI, GTC China 2017 - ID 100
Streaming:
Share:
Intelligent Machines and IoT
Presentation
Media
Keynote Address
Opening Keynote Speech ...Read More

Opening Keynote Speech

  Back
 
Keywords:
Intelligent Machines and IoT, GTC Washington D.C. 2016 - ID DCS16158
Streaming:
Share:
Federal
Presentation
Media
Advancing the Frontiers of Science
The National Science Foundation (NSF) is an independent federal agency that supports fundamental research and education across all fields of science and engineering. With an annual budget of $7.5 billion, NSF awards grants to nearly 2,000 colleg ...Read More

The National Science Foundation (NSF) is an independent federal agency that supports fundamental research and education across all fields of science and engineering. With an annual budget of $7.5 billion, NSF awards grants to nearly 2,000 colleges, universities and other institutions in all 50 states. Hear how NSF is advancing discovery and technological innovation in all fields, including artificial intelligence, to keep the United States at the forefront of global science and engineering leadership.

  Back
 
Keywords:
Federal, GTC Washington D.C. 2016 - ID DCS16164
Streaming:
Share:
Healthcare and Life Sciences
Presentation
Media
Cancer Research and Technology Cancer Moonshot Project
In this keynote, we'll show how the Cancer Moonshot Task Force under Vice President Biden is unleashing the power of data to help end cancer as we know it. We'll discuss global efforts inspired by the Cancer Moonshot that will empower A. ...Read More

In this keynote, we'll show how the Cancer Moonshot Task Force under Vice President Biden is unleashing the power of data to help end cancer as we know it. We'll discuss global efforts inspired by the Cancer Moonshot that will empower A.I. and deep learning for oncology with larger and more accessible datasets.

  Back
 
Keywords:
Healthcare and Life Sciences, GTC Washington D.C. 2016 - ID DCS16165
Streaming:
Share:
Intelligent Machines and IoT
Presentation
Media
The Economic Implications of Artificial Intelligence
 
Keywords:
Intelligent Machines and IoT, GTC Washington D.C. 2016 - ID DCS16182
Streaming:
Share:
Deep Learning and AI
Presentation
Media
Opening Keynote.
Opening Keynote. ...Read More

Opening Keynote.

  Back
 
Keywords:
Deep Learning and AI, AI Conference Australia 2016 - ID AUS6123
Streaming:
Share:
 
Customer Keynote
What New Results in Visual Question Answering Have to Say about Old AI ...Read More

What New Results in Visual Question Answering Have to Say about Old AI

  Back
 
Keywords:
Deep Learning and AI, AI Conference Australia 2016 - ID AUS6124
Streaming:
Share:
 
Welcome Speech
Opening Keynote Speech ...Read More

Opening Keynote Speech

  Back
 
Keywords:
Deep Learning and AI, AI Conference Australia 2016 - ID AUS6126
Streaming:
Share:
Virtual Reality and Augmented Reality
Presentation
Media
NVIDIA Keynote, CEO
 
Keywords:
Virtual Reality and Augmented Reality, Deep Learning and AI, Algorithms and Numerical Techniques, AI Startup, GTC Japan 2016 - ID 1000
Streaming:
Download:
Share:
Machine Learning & Deep Learning
Presentation
Media
Opening Keynote
Don't miss GTC's opening keynote address from NVIDIA CEO and co-founder Jensen Huang. He'll discuss the latest breakthroughs in visual computing, including how NVIDIA is fueling the revolution in deep learning. ...Read More

Don't miss GTC's opening keynote address from NVIDIA CEO and co-founder Jensen Huang. He'll discuss the latest breakthroughs in visual computing, including how NVIDIA is fueling the revolution in deep learning.

  Back
 
Keywords:
Machine Learning & Deep Learning, GTC Silicon Valley 2015 - ID S2000
Streaming:
Download:
Share:
 
Large-Scale Deep Learning For Building Intelligent Computer Systems
Over the past few years, we have built large-scale computer systems for training neural networks, and then applied these systems to a wide variety of problems that have traditionally been very difficult for computers. We have made significant im ...Read More

Over the past few years, we have built large-scale computer systems for training neural networks, and then applied these systems to a wide variety of problems that have traditionally been very difficult for computers. We have made significant improvements in the state-of-the-art in many of these areas, and our software systems and algorithms have been used by dozens of different groups at Google to train state-of-the-art models for speech recognition, image recognition, various visual detection tasks, language modeling, language translation, and many other tasks. In this talk, I''ll highlight some of the distributed systems and algorithms that we use in order to train large models quickly. I''ll then discuss ways in which we have applied this work to a variety of problems in Google''s products, usually in close collaboration with other teams. This talk describes joint work with many people at Google.

  Back
 
Keywords:
Machine Learning & Deep Learning, GTC Silicon Valley 2015 - ID S5817
Streaming:
Download:
Share:
 
Deep Learning: What's Next
Deep Learning has transformed many important tasks, including speech and image recognition. Deep Learning systems scale well by absorbing huge amounts of data to create accurate models. The computational resources afforded by GPUs have been instrumen ...Read More
Deep Learning has transformed many important tasks, including speech and image recognition. Deep Learning systems scale well by absorbing huge amounts of data to create accurate models. The computational resources afforded by GPUs have been instrumental to this scaling. However, as Deep Learning has become more mainstream, it has generated some hype, and has been linked to everything from world peace to evil killer robots. In this talk, Dr. Ng will help separate hype from reality, and discuss potential ways that Deep Learning technologies can benefit society in the short and long term.  Back
 
Keywords:
Machine Learning & Deep Learning, Computer Vision and Machine Vision, GTC Silicon Valley 2015 - ID S5818
Streaming:
Download:
Share:
Keynote
Presentation
Media
Opening Keynote
Don''t miss the opening keynote feature Jensen Huang, Co-Founder, President, and CEO of NVIDIA. Hear about what''s next in visual computing, and preview disruptive technologies and exciting demonstrations across industries. ...Read More

Don''t miss the opening keynote feature Jensen Huang, Co-Founder, President, and CEO of NVIDIA. Hear about what''s next in visual computing, and preview disruptive technologies and exciting demonstrations across industries.

  Back
 
Keywords:
Keynote, General Interest, GTC Silicon Valley 2014 - ID S4000
Streaming:
Share:
 
Keynote: Video Games and the Future of Cognitive Enhancement
A fundamental challenge of modern society is the development of effective approaches to enhance brain function and cognition in both healthy and impaired individuals. For the healthy, this serves as a core mission of our educational system and f ...Read More

A fundamental challenge of modern society is the development of effective approaches to enhance brain function and cognition in both healthy and impaired individuals. For the healthy, this serves as a core mission of our educational system and for the cognitively impaired this is a critical goal of our medical system. Unfortunately, there are serious and growing concerns about the ability of either system to meet this challenge. I will describe an approach developed in our lab that uses custom-designed video games to achieve meaningful and sustainable cognitive enhancement (e.g., Anguera, et al. Nature 2013), as well the next stage of our research program, which uses video games integrated with technological innovations in software (e.g., brain computer interface algorithms, GPU computing) and hardware (e.g., virtual reality headsets, mobile EEG, transcranial electrical brain stimulation) to create a novel personalized closed loop system. I will share with you a vision of the future in which high-tech is used as an engine to enhance our brain''s information processing systems, thus reducing our reliance on non-specific drugs to treat neurological and psychiatric conditions and allowing us to better target our educational efforts. This keynote will be preceded by naming the winner of the CUDA Center of Excellence Achievement Award, winner for Best Poster, and the new CUDA Fellows, followed by the launch announcement of the Global Impact Award. (Award ceremony duration approximately 15 minutes).

  Back
 
Keywords:
Keynote, Medical Imaging and Radiology, Video and Image Processing, GTC Silicon Valley 2014 - ID S4780
Streaming:
Download:
Share:
 
Keynote: Using NVIDIA GPUs for Feature Film Production at Pixar
This presentation will show how Pixar uses GPU technology to empower artists in the animation and lighting departments. By providing our artists with high-quality, interactive visual feedback, we enable them to spend more time making creative de ...Read More

This presentation will show how Pixar uses GPU technology to empower artists in the animation and lighting departments. By providing our artists with high-quality, interactive visual feedback, we enable them to spend more time making creative decisions. Animators interactively pose characters in order to create a performance. When features like displacement, fur, and shadows become critical for communicating the story, it is vital to be able to represent these visual elements in motion at interactive frame rates. We will show Presto, Pixar''s proprietary animation system, which uses GPU acceleration to deliver real-time feedback during the character animation process, using examples from Pixar''s recent films. Lighting artists place and adjust virtual lights to create the mood and tone of the scene as well as guide the audience''s attention. A physically-based illumination model allows these artists to create visually-rich imagery using simpler and more direct controls. We will demonstrate our interactive lighting preview tool, based on this model, built on NVIDIA''s OptiX framework, and fully integrated into our new Katana-based production workflow.

  Back
 
Keywords:
Keynote, Media and Entertainment, GTC Silicon Valley 2014 - ID S4884
Streaming:
Share:
General Interest
Presentation
Media
GTC Keynote with Jensen Huang, NVIDIA
Don''t miss the opening keynote feature Jensen Huang, Co-Founder, President, and CEO of NVIDIA. Hear about what''s next in computing and graphics, and preview disruptive technologies and exciting demonstrations across industries. ...Read More

Don''t miss the opening keynote feature Jensen Huang, Co-Founder, President, and CEO of NVIDIA. Hear about what''s next in computing and graphics, and preview disruptive technologies and exciting demonstrations across industries.

  Back
 
Keywords:
General Interest, GTC Silicon Valley 2013 - ID S3900
Streaming:
Download:
Share:
 
GTC Keynote featuring Erez Lieberman Aiden of Baylor College of Medicine and Rice University - Parallel Processing of the Genomes, by the Genomes and for the Genomes
The human genome is a sequence of 3 billion chemical letters inscribed in a molecule called DNA. Famously, short stretches (~10 letters, or a-base pairs) of DNA fold into a double helix. But what about longer pieces? How does a 2 meter long macr ...Read More

The human genome is a sequence of 3 billion chemical letters inscribed in a molecule called DNA. Famously, short stretches (~10 letters, or a-base pairs) of DNA fold into a double helix. But what about longer pieces? How does a 2 meter long macromolecule, the genome, fold up inside a 6 micrometer wide nucleus? And, once packed, how does the information contained in this ultra-dense structure remain accessible to the cell? This talk will discuss how the human genome folds in three dimensions, a folding enables the cell to access and process massive quantities of information in parallel. To probe how genomes fold, we developed Hi-C, together with collaborators at the Broad Institute and UMass Medical School. Hi-C couples proximity-dependent DNA ligation and massively parallel sequencing. To analyze our data and reconstruct the underlying folds, we, too must engage in massively parallel computation. I will describe how we use NVIDIA's CUDA technology to analyze our results and simulate the physical processes of genome folding and unfolding.

  Back
 
Keywords:
General Interest, Developer - Algorithms, Bioinformatics & Genomics, GTC Silicon Valley 2013 - ID S3901
Streaming:
Share:
 
GTC Keynote featuring Ralph Gilles of Chrysler
Ralph Gilles, senior vice president of Product Design and president and CEO of SRT (Street and Racing Technology) Brand and Motorsports at Chrysler Group LLC and the mind behind some of the company most innovative products, will provide a behind ...Read More

Ralph Gilles, senior vice president of Product Design and president and CEO of SRT (Street and Racing Technology) Brand and Motorsports at Chrysler Group LLC and the mind behind some of the company most innovative products, will provide a behind-the-scenes look at the auto industry. Gilles will review how GPUs are used to advance every step of the automobile development process from the initial conceptual designs and engineering phases through product assembly and marketing. He will also discuss and how Chrysler Group utilizes GPUs and the latest technologies to build better, safer cars and reduce time to market.

  Back
 
Keywords:
General Interest, Automotive, GTC Silicon Valley 2013 - ID S3902
Streaming:
Share:
 
Opening Keynote
Do not miss the opening keynote, featuring Jensen Huang, CEO and Co-Founder of NVIDIA. Hear about what's next in computing and graphics, and preview disruptive technologies and exciting demonstrations from across industries. Jen-Hsun co-foun ...Read More

Do not miss the opening keynote, featuring Jensen Huang, CEO and Co-Founder of NVIDIA. Hear about what's next in computing and graphics, and preview disruptive technologies and exciting demonstrations from across industries. Jen-Hsun co-founded NVIDIA in 1993 and has served since its inception as president, chief executive officer and a member of the board of directors.

  Back
 
Keywords:
General Interest, GTC Silicon Valley 2012 - ID S2000
Streaming:
Share:
 
Day 2 Keynote: From Democratic Consensus to Cannibalistic Hordes: GPU Computing Reveals the Principles of Collective Behavior
Collective behavior is one of the most pervasive features of the natural world. Our brains are composed of billions of interconnected cells communicating with chemical and electrical signals. We are integrated in our own human society. Elsewhere ...Read More

Collective behavior is one of the most pervasive features of the natural world. Our brains are composed of billions of interconnected cells communicating with chemical and electrical signals. We are integrated in our own human society. Elsewhere in the natural world a fish school convulses, as if one entity, when being attacked by a predator. How does individual behavior produce dynamic group-level properties? Do animal groups -or even cells in a tumor- function as some form of collective mind? How does socially contagious behavior spread through natural human crowds? In his keynote address, Prof. Iain D. Couzin, Professor of Ecology and Evolutionary Biology at Princeton University, will demonstrate how GPU computing has been pivotal in the study of collective behavior, helping reveal how collective action emerges in a wide range of groups from plague locusts to human crowds, and the critical role that uninformed, or weakly-opinionated, individuals play in democratic consensus decision-making.

  Back
 
Keywords:
General Interest, GTC Silicon Valley 2012 - ID S2001
Streaming:
Share:
 
Day 3 Keynote: Not Your Grandfather's Moon Landing
Do not miss the day 3 keynote, featuring Part-Time Scientists Robert Boehme and Wes Faler. Boehme and Faler are part of a team of international scientists and engineers who want to send a rover to the moon before the end of the year 2013. In thi ...Read More

Do not miss the day 3 keynote, featuring Part-Time Scientists Robert Boehme and Wes Faler. Boehme and Faler are part of a team of international scientists and engineers who want to send a rover to the moon before the end of the year 2013. In this presentation, they will discuss their goals, recent accomplishments and milestones, and how GPUs have help in unexpected ways.

  Back
 
Keywords:
General Interest, GTC Silicon Valley 2012 - ID S3002
Streaming:
Share:
 
Opening Keynote with Jensen Huang, NVIDIA
Jensen Huang
Do not miss this opening keynote, featuring Jensen Huang, CEO and Co-Founder of NVIDIA and special guests. Hear about what’s next in gpu computing, and preview disruptive technologies and exciting demonstrations from across industries. Jen ...Read More

Do not miss this opening keynote, featuring Jensen Huang, CEO and Co-Founder of NVIDIA and special guests. Hear about what’s next in gpu computing, and preview disruptive technologies and exciting demonstrations from across industries. Jensen Huang co-founded NVIDIA in 1993 and has served since its inception as president, chief executive officer and a member of the board of directors.

  Back
 
Keywords:
General Interest, GTC China 2011 - ID S10001
Download:
Share:
HPC and AI
Presentation
Media
The Evolution of Modern Parallel Computing
Sanford Russell
- NVIDIA
 
Keywords:
HPC and AI, GTC Taiwan 2011 - ID GTCT1101
Download:
Share:
General Interest
Presentation
Media
Opening Keynote with Jensen Huang, NVIDIA
Jensen Huang
The opening keynote, features Jensen Huang, CEO and Co-Founder of NVIDIA and special guests. Hear about what''s next in computing and graphics, and preview disruptive technologies and exciting demonstrations from across industries. ...Read More

The opening keynote, features Jensen Huang, CEO and Co-Founder of NVIDIA and special guests. Hear about what''s next in computing and graphics, and preview disruptive technologies and exciting demonstrations from across industries.

  Back
 
Keywords:
General Interest, GTC Silicon Valley 2010 - ID S091001
Download:
Share:
 
Day 2 Keynote with Dr. Klaus Schluten, University of Illinois at Urbana-Champaign
Klaus Schluten
How does the H1N1 "Swine Flu" virus avoid drugs while attacking our cells? What can we learn about solar energy by studying biological photosynthesis? How do our cells read the genetic code? What comes next in computational biology? Co ...Read More

How does the H1N1 "Swine Flu" virus avoid drugs while attacking our cells? What can we learn about solar energy by studying biological photosynthesis? How do our cells read the genetic code? What comes next in computational biology? Computational biology is approaching a new and exciting frontier: the ability to simulate structures and processes in living cells. Come learn about the "computational microscope," a new research instrument that scientists can use to simulate biomolecules at nearly infinite resolution. The computational microscope complements the most advanced physical microscopes to guide today's biomedical research. In this keynote address, computational biology pioneer Dr. Klaus Schulten of the University of Illinois, Urbana-Champaign, will introduce the computational microscope, showcase the widely used software underlying it, and highlight major discoveries made with the aid of the computational microscope ranging from viewing protein folding, translating the genetic code in cells, and harvesting solar energy in photosynthesis. He will also look towards a future when cell tomography and computing will establish atom-by-atom views of entire life forms.

  Back
 
Keywords:
General Interest, Life & Material Science, GTC Silicon Valley 2010 - ID S10002
Download:
Share:
 
Closing Keynote with Dr. Sebastien Thrun, Stanford University and Google
Sebastien Thrun
What really causes accidents and congestion on our roadways? How close are we to fully autonomous cars? In his keynote address, Stanford Professor and Google Distinguished Engineer, Dr. Sebastian Thrun, will show how his two autonomous vehicles, ...Read More

What really causes accidents and congestion on our roadways? How close are we to fully autonomous cars? In his keynote address, Stanford Professor and Google Distinguished Engineer, Dr. Sebastian Thrun, will show how his two autonomous vehicles, Stanley (DARPA Grand Challenge winner), and Junior (2nd Place in the DARPA Urban Challenge) demonstrate how close yet how far away we are to fully autonomous cars. Using computer vision combined with lasers, radars, GPS sensors, gyros, accelerometers, and wheel velocity, the vehicle control systems are able to perceive and plan the routes to safely navigate Stanley and Junior through the courses. However, these closed courses are a far cry from everyday driving. Find out what the team will do next to get one step closer to the "holy grail" of computer vision, and a huge leap forward toward the concept of fully autonomous vehicles.

  Back
 
Keywords:
General Interest, Computer Vision and Machine Vision, Machine Learning & Deep Learning, GTC Silicon Valley 2010 - ID S10003
Download:
Share:
Medical Imaging and Radiology
Presentation
Media
Graphcuts with CUDA and Applications in Image Processing
Timo Stich
Graph Cuts is a powerful and popular optimization approach to solve image processing problems such as image segmentation, stereo vision, image restoration and many more. ...Read More
Graph Cuts is a powerful and popular optimization approach to solve image processing problems such as image segmentation, stereo vision, image restoration and many more. In this talk, we present CUDA implementations of the push-relabel algorithm to compute Graph Cuts. Starting from the basic algorithm we discuss its parallel processing properties. Then different optimization strategies are explored and their strengths and weaknesses are evaluated. We conclude by exploring applications of Graph Cuts to solve image processing problems using GPUs.  Back
 
Keywords:
Medical Imaging and Radiology, Medical Imaging and Radiology, GTC Silicon Valley 2009 - ID S91060
Streaming:
Download:
Share:
Visualization
Presentation
Media
4D Volume Rendering
Shalini Venkataraman
With advances in image acquisition and numerical simulation techniques, fields ranging from medical imaging to astrophysics are producing data with very large spatial and temporal resolutions. ...Read More
With advances in image acquisition and numerical simulation techniques, fields ranging from medical imaging to astrophysics are producing data with very large spatial and temporal resolutions. Interactive visualization techniques are crucial to understand and isolate features from the resulting large time dependent 4D volumetric data. This presentation explores the various rendering methods such as texture slicing, raycasting in graphics and cuda as well as hybrid approaches showing their promises and pitfalls. It is common for 4D data to exceed the graphics memory capabilities and approaches for efficiently streaming data such as PBO's and CPU/GPU asynchronous modes are explained. We conclude with a discussion on how other related solutions from NVIDIA can be integrated, specifically focusing on 3D Vision stereo and NVScale middleware to harness multiple GPU's for distributed rendering.  Back
 
Keywords:
Visualization, Energy Exploration, Film, Medical Imaging and Radiology, Visualization, GTC Silicon Valley 2009 - ID S91102
Download:
Share:
Film
Presentation
Media
Strategies for GPU Acceleration of Common Visual Effects for Broadcast and Post-Production
Since 1995 BorisFX has developed image processing and 3D graphics software for Broadcast and Post-Production, with a particular focus on staple visual effects such as 3D Text, Chroma Key, and Film Look. ...Read More
Since 1995 BorisFX has developed image processing and 3D graphics software for Broadcast and Post-Production, with a particular focus on staple visual effects such as 3D Text, Chroma Key, and Film Look. While achieving award-winning quality, these CPU bound effects have lacked the interactivity today''s customers demand. With the advance of GPU hardware from NVIDIA the ability to accelerate these effects has become a reality, and allowed for many previously impractical features to present themselves. In this presentation we will demonstrate how the use of the GPU has benefited our products in terms of performance and features in our creation of GPU-Rendered 3D Text, and discuss the strategies we employed to emphasize the benefits and minimize the drawbacks of the GPU in building the multi-pass Chroma Key and Film Look filters.   Back
 
Keywords:
Film, Film, Visualization, GTC Silicon Valley 2009 - ID S91114
Download:
Share:
General Interest
Presentation
Media
Day 2 Keynote with Hanspeter Pfister, Harvard University
Hanspeter Pfister
High-Throughput Science How did the universe start? How is the brain wired? How does matter interact at the quantum level? These are some of the great scientific challenges of our times, and answering them requires bigger scientific instruments, ...Read More

High-Throughput Science How did the universe start? How is the brain wired? How does matter interact at the quantum level? These are some of the great scientific challenges of our times, and answering them requires bigger scientific instruments, increasingly precise imaging equipment and ever-more complex computer simulations. In his keynote address, Harvard professor, researcher and computing visionary Hanspeter Pfister will discuss the computational obstacles scientists face and how commodity high-throughput computing can enable high-throughput science, in which massive data streams are processed and analyzed rapidly -- from the instrument through to the desktop. Finally Professor Pfister will survey several groundbreaking projects at Harvard that leverage GPUs for high- throughput science, ranging from radio astronomy and neuroscience to quantum chemistry and physics.

  Back
 
Keywords:
General Interest, GTC Silicon Valley 2009 - ID S91422
Streaming:
Download:
Share:
 
Day 3 Keynote with Richard Kerris, Lucasfilm
Richard Kerris
Games and interactive media have long been the beneficiaries of cutting edge GPU technology and it has not gone unnoticed in the world of feature film production. To date the visual effects industry had been a sideline observer of these advances ...Read More

Games and interactive media have long been the beneficiaries of cutting edge GPU technology and it has not gone unnoticed in the world of feature film production. To date the visual effects industry had been a sideline observer of these advances while awaiting technology to reach maturity. At Lucasfilm, research and development has been on-going for some time and this past summer Industrial Light & Magic employed this technology in two of its summer blockbuster films. Lucasfilm CTO, Richard Kerris, will show a brief history of their computer graphics for film, and will then pull back the curtain on how they are now using GPU technology to advance the state of the art in visual effects and provide a glimpse of what's on the horizon for GPU's in future and how it will impact filmmaking.

  Back
 
Keywords:
General Interest, GTC Silicon Valley 2009 - ID S91423
Streaming:
Share:
AEC Industries
Presentation
Media
An Architectural Design Firm's Journey Through Virtual GPU Technology for Global Collaboration
Learn the benefits that virtualization provides for an architecture and engineering design firm, along with the journey through the advancements in virtualization technology it took to finally meet the graphics-intensive needs of our design software. ...Read More
Learn the benefits that virtualization provides for an architecture and engineering design firm, along with the journey through the advancements in virtualization technology it took to finally meet the graphics-intensive needs of our design software. We'll share our experiences in how virtualization allows a large company, with over 15 offices and 1,000 people worldwide, to collaborate and work as a single firm. We'll show some cost comparisons with virtualization, along with their management benefits and requirements. We'll also look at the methods we used to set and test metrics specific to our requirements, and follow the results of those metrics through the changes in graphics virtualization technology.  Back
 
Keywords:
AEC Industries, GPU Virtualization, Data Center and Cloud Infrastructure, GTC Silicon Valley 2017 - ID S7174
Download:
Share:
 
Scalable Enterprise Visualization
We'll discuss Bunsen, a large-scale visualization framework that prepares and optimizes engineering, architectural, and other CAD and CAM data. Bunsen is a cloud-hosted solution that reads and writes various industry standard file formats (for examp ...Read More
We'll discuss Bunsen, a large-scale visualization framework that prepares and optimizes engineering, architectural, and other CAD and CAM data. Bunsen is a cloud-hosted solution that reads and writes various industry standard file formats (for example, Revit, SOLIDWORKS, Rhino, Maya, Max, Siemens, and Microstation) and provides powerful tools for processing and conversion. It runs on public cloud solutions, such as AWS or Google, or within your own data center or on-prem cloud. All hardware and software are provisioned in the cloud and are usable from any laptop, tablet, or phone with a web browser. Within Bunsen, the user can create sets of reusable rules to process data for visualization and output. You can think of these rules as company standards relating to lighting, materials, colors, and how to reduce object complexity. Possible visualization output platforms include rendering and animation, virtual reality, augmented reality, and real-time game engines, such as Unreal and Unity. Bunsen doesn't mean you change your workflow -- it is a framework to automate, document, and accelerate your existing workflows.  Back
 
Keywords:
AEC Industries, Manufacturing Industries, Rendering and Ray Tracing, GTC Silicon Valley 2017 - ID S7474
Download:
Share:
 
From Cracks to Hard Hats: Focusing on Industrial Computer Vision
We'll present, in a case study driven presentation, specific examples of how GPU-enabled deep neural networks are powering new methods for analyzing the content of photos and videos from industrial contexts. First, we'll present a collaboration bet ...Read More
We'll present, in a case study driven presentation, specific examples of how GPU-enabled deep neural networks are powering new methods for analyzing the content of photos and videos from industrial contexts. First, we'll present a collaboration between Smartvid.io and Engineering News-Record, the leading publication in the architecture, engineering, and construction vertical. This ongoing initiative leverages computer vision techniques and semantic approaches to help identify and indicate safe and unsafe situations in jobsite photos. Second, we'll present a collaboration with Arup, a London-based engineering firm, on the use of specific classifiers to localize and measure cracks and related defects in infrastructure.  Back
 
Keywords:
AEC Industries, Deep Learning and AI, AI Startup, GTC Silicon Valley 2017 - ID S7575
Download:
Share:
 
Design with Virtual Reality in Architecture, Engineering and Construction
Learn how Gensler is using the latest technology in virtual reality across all aspects of the design process for the AEC industry. We'll cover how VR has added value to the process when using different kinds of VR solutions. Plus we'll t ...Read More

Learn how Gensler is using the latest technology in virtual reality across all aspects of the design process for the AEC industry. We'll cover how VR has added value to the process when using different kinds of VR solutions. Plus we'll talk about some of the challenges Gensler has faced with VR in terms of hardware, software, and workflows. Along with all of this, NVIDIA's latest VR visualization tools are helping with the overall process and realism of our designs.

  Back
 
Keywords:
AEC Industries, Virtual Reality and Augmented Reality, GTC Silicon Valley 2017 - ID S7614
Download:
Share:
 
利用大型 AEC 场景、先进的逼真材质和 VR 作出设计决策
大型 AEC 项目涉及复杂结构的设计与验证工作,但也需通过 VR 来让用户身临其境地感受和理解建筑空间的实际大小。SHoP Architects 是 Dassault Systèmes 3DEXPERIENCE 的早期采用者之一,该平台无需使用其他外部工具便可将 CATIA CAD 领域 20 多年积累的优秀成果与先进的渲染材质(包括 Substance 支持)结合在一起,并可提供原生 VR 沉浸式体验。本次报告将展示 SHoP 主要借助 GPU 实现的最新成果,包括如何使用上述平台模拟建筑施工、通过 VR SLI 以 90 FPS 帧速率渲染大型网格以显著提高性能,从而实现多用户多地点 VR 评审,以及通过集成 Substance 来开展逼真的 AEC 设计评审 ...Read More
大型 AEC 项目涉及复杂结构的设计与验证工作,但也需通过 VR 来让用户身临其境地感受和理解建筑空间的实际大小。SHoP Architects 是 Dassault Systèmes 3DEXPERIENCE 的早期采用者之一,该平台无需使用其他外部工具便可将 CATIA CAD 领域 20 多年积累的优秀成果与先进的渲染材质(包括 Substance 支持)结合在一起,并可提供原生 VR 沉浸式体验。本次报告将展示 SHoP 主要借助 GPU 实现的最新成果,包括如何使用上述平台模拟建筑施工、通过 VR SLI 以 90 FPS 帧速率渲染大型网格以显著提高性能,从而实现多用户多地点 VR 评审,以及通过集成 Substance 来开展逼真的 AEC 设计评审。  Back
 
Keywords:
AEC Industries, VR and Simulation, GTC China 2018 - ID CH81001
Share:
 
前沿边缘计算技术助力重新构想未来城市规划
新型视觉计算技术正在改变建筑和城市的设计方式。无论是在设计评审和客户展示阶段,还是在设计流程的早期阶段,设计公司都在更广泛地应用照片级逼真度,以帮助改进设计决策。虚拟现实有助我们更加清晰地了解设计评审,而机器学习和深度学习让图像分析、预测和自然语言处理在工程应用中得以实现。本次讲座将回顾这些突破性技术如何促进设计团队实现创新,并将涵盖在工程领域部署和推广机器学习的趋势与挑战 ...Read More
新型视觉计算技术正在改变建筑和城市的设计方式。无论是在设计评审和客户展示阶段,还是在设计流程的早期阶段,设计公司都在更广泛地应用照片级逼真度,以帮助改进设计决策。虚拟现实有助我们更加清晰地了解设计评审,而机器学习和深度学习让图像分析、预测和自然语言处理在工程应用中得以实现。本次讲座将回顾这些突破性技术如何促进设计团队实现创新,并将涵盖在工程领域部署和推广机器学习的趋势与挑战。  Back
 
Keywords:
AEC Industries, GTC China 2018 - ID CH81002
Download:
Share:
AI Application Deployment and Inference
Presentation
Media
Deep Learning Implementers Panel: Field Insights for Accelerating Deep Learning Performance, Productivity and Scale
This customer panel brings together A.I. implementers who have deployed deep learning at scale using NVIDIA DGX Systems. We'll focus on specific technical challenges we faced, solution design considerations, and best practices learned from i ...Read More

This customer panel brings together A.I. implementers who have deployed deep learning at scale using NVIDIA DGX Systems. We'll focus on specific technical challenges we faced, solution design considerations, and best practices learned from implementing our respective solutions. Attendees will gain insights such as: 1) how to set up your deep learning project for success by matching the right hardware and software platform options to your use case and operational needs; 2) how to design your architecture to overcome unnecessary bottlenecks that inhibit scalable training performance; and 3) how to build an end-to-end deep learning workflow that enables productive experimentation, training at scale, and model refinement.

  Back
 
Keywords:
AI Application Deployment and Inference, AI and DL Business Track (high level), Data Center and Cloud Infrastructure, AI for Business, HPC and Supercomputing, GTC Silicon Valley 2018 - ID S8194
Streaming:
Download:
Share:
 
Deploying Deep Neural Networks as a Service Using TensorRT and NVIDIA-Docker
Learn how you can utilize TensorRT and NVIDIA Docker to quickly configure and deploy a GPU-accelerated inference server and start gaining insights from your trained deep neural network (DNN) models. TensorRT is a high-performance tool for low-latency ...Read More
Learn how you can utilize TensorRT and NVIDIA Docker to quickly configure and deploy a GPU-accelerated inference server and start gaining insights from your trained deep neural network (DNN) models. TensorRT is a high-performance tool for low-latency, high-throughput DNN inference. The latest release of TensorRT introduces a novel, framework-agnostic network definition format called universal framework format, which allows TensorRT to support and optimize DNN models trained in multiple deep learning frameworks. We'll leverage the TensorRT Python API to create a lightweight Python Flask application capable of serving multiple DNN models trained using TensorFlow, PyTorch, and Caffe, and also discuss how to containerize this inference service using NVIDIA Docker for ease of deployment at scale. This session will consist of a lecture, live demos, and detailed instructions.  Back
 
Keywords:
AI Application Deployment and Inference, Tools and Libraries, Data Center and Cloud Infrastructure, GTC Silicon Valley 2018 - ID S8495
Streaming:
Download:
Share:
 
Monte Carlo Methods and Neural Networks
The average human brain has about 100 billion nerve cells. We therefore investigate the question whether there are algorithms for artificial neural networks that are linear in the number of neurons, while the number of connections incident to a neuro ...Read More
The average human brain has about 100 billion nerve cells. We therefore investigate the question whether there are algorithms for artificial neural networks that are linear in the number of neurons, while the number of connections incident to a neuron is bounded by a constant. We offer two approaches to answer this question: First, we derive an algorithm that quantizes a trained artificial neural network such that the resulting complexity is linear. Second, we demonstrate that training networks, whose connections are determined by uniform sampling can achieve a similar precision as compared to using fully connected layers. Due to sparsity upfront, these networks can be trained much faster. Both approaches are made plausible by relating artificial neural units to Monte Carlo integration. We'll demonstrate the results for classic test datasets.  Back
 
Keywords:
AI Application Deployment and Inference, AI and DL Research, GTC Silicon Valley 2018 - ID S8780
Streaming:
Download:
Share:
 
AI Solutions and Use Cases Up Close (Presented by Inspur Systems)
Inspur has been deploying AI solutions with our customers, such as Microsoft, Alibaba, Baidu, BMW, for many years. We will share AI use cases on how we deploy AI at scale and take a close look at the technologies that enable AI deployments.
Inspur has been deploying AI solutions with our customers, such as Microsoft, Alibaba, Baidu, BMW, for many years. We will share AI use cases on how we deploy AI at scale and take a close look at the technologies that enable AI deployments.  Back
 
Keywords:
AI Application Deployment and Inference, AI and DL Research, HPC and AI, GTC Silicon Valley 2018 - ID S8996
Streaming:
Share:
 
Putting AI to Work in an Enterprise: Deep Learning as a Service (Presented by IBM)
Now that Deep learning has moved out of the lab and into production, how do you provide training environments to all your internal customers working across business units with different requirements and avoid provisioning separate clusters? IBM has a ...Read More
Now that Deep learning has moved out of the lab and into production, how do you provide training environments to all your internal customers working across business units with different requirements and avoid provisioning separate clusters? IBM has applied decades of HPC experience to build a production ready learning stack, including servers accelerated with NVIDIA GPUs, workload and resource management software, ready to use open source frameworks and it's all covered by IBM support. The solution provides a secure multi-tenant environment so multiple data scientists can share a common set of resources, eliminating silos, while running multiple instances of the same or different applications. The deep learning effort is enhanced with end-to-end pipeline support from data ingestion and preparation, through model training and tuning, to inference. In this session, we will explore what an enterprise deep learning environment looks like and provide insights into the unique IBM value for accelerating the use of deep learning across a wide variety of industries.  Back
 
Keywords:
AI Application Deployment and Inference, GTC Silicon Valley 2018 - ID S81049
Streaming:
Download:
Share:
 
GPU-Powered Megacity Scale Transport Management, Municipal Services and Public Safety Solutions
Learn how VisionLabs GPU-powered solutions contribute to creating a safer, smarter Megacity a metropolitan area with a total population in excess of ten million people. We'll do a deep dive into three implemented and ongoing huge scale smart-city ...Read More
Learn how VisionLabs GPU-powered solutions contribute to creating a safer, smarter Megacity a metropolitan area with a total population in excess of ten million people. We'll do a deep dive into three implemented and ongoing huge scale smart-city projects, understand challenges, technical specifics and how GPU computing impacts each of these cases: Face authentication-based immobilizer and driver monitoring systems for municipal service vehicles powered by the NVIDIA Jetson TX2 embedded platform; Megacity scale vehicle traffic analysis and anomalies detection powered by NVIDIA Tesla P40 with over 80 million daily recognition requests; National scale face identification platform for financial services with over 110 million faces in its database. The foundation of all these projects is VisionLabs LUNA a cross-platform object recognition software based on proprietary deep neural networks (DNN) inference framework. To build cost-effective solutions, VisionLabs use know-hows in DNN quantization and acceleration. In terms of accuracy, VisionLabs is recognized as a top three best in the world by National Institute of Standards and Technology's face recognition vendor test, and LFW by University of Massachusetts challenges.  Back
 
Keywords:
AI Application Deployment and Inference, NVIDIA Inception Program, Intelligent Video Analytics and Smart Cities, Deep Learning and AI Frameworks, Computer Vision, GTC Silicon Valley 2018 - ID S8584
Streaming:
Download:
Share:
 
VACnet: Using Deep Learning to Combat Cheating in 'Counter-Strike: Global Offensive'
We'll delve into the nuts and bolts of how Valve has utilized deep learning to combat cheating in "Counter-Strike: Global Offensive." We'll cover total system details, from the high-level server architecture to the low-level features fed ...Read More
We'll delve into the nuts and bolts of how Valve has utilized deep learning to combat cheating in "Counter-Strike: Global Offensive." We'll cover total system details, from the high-level server architecture to the low-level features fed into the AI. Deep learning has proven to be very effective at identifying cheating behavior without any client-side instrumentation, making it robust against malicious attack by cheaters and cheat vendors. By retraining regularly, the network continues to evolve, picking up new cheating behaviors within hours of their appearance. As a result of this approach, certain types of cheats have been reduced by a factor of 100.  Back
 
Keywords:
AI Application Deployment and Inference, AI for Gaming, GTC Silicon Valley 2018 - ID S8732
Streaming:
Share:
 
Autoregressive Wavenet Inference on Volta GPUs
Autoregressive wavenets have demonstrated extremely high quality real-time speech synthesis results.  However, the compute requirements and tight latency bounds have made them impractical for deployment on traditional CPU-only systems.  In ...Read More
Autoregressive wavenets have demonstrated extremely high quality real-time speech synthesis results.  However, the compute requirements and tight latency bounds have made them impractical for deployment on traditional CPU-only systems.  In this talk we demonstrate that Volta GPUs provide excellent real-time inference performance on these networks, making practical deployments possible.  We discuss several alternative implementation techniques and demonstrate their achieved performance on a V100 GPU.  Back
 
Keywords:
AI Application Deployment and Inference, Speech and Language Processing, GTC Silicon Valley 2018 - ID S8968
Streaming:
Share:
 
Adopting Artificial Intelligence Technologies in Networking (Presented by Cisco)
This talk will provide an overview of what is happening in the world of artificial intelligence as it relates to networking, IT infrastructure, and IoT technologies. We will broadly cover AI topics ranging from machine learning and deep learning to s ...Read More
This talk will provide an overview of what is happening in the world of artificial intelligence as it relates to networking, IT infrastructure, and IoT technologies. We will broadly cover AI topics ranging from machine learning and deep learning to symbolic AI. Applied AI as well as general AI and their hybrids are all critical in solving many of today's complex long tail problems in real-time. Just as the capabilities, business opportunities, and positive benefits of AI are growing at a seemingly exponential rate so are the security vulnerabilities, failure modes, and potential adverse business impacts. We will discuss new hybrid neural symbolic approaches that promise to address these issues while simultaneously opening the door to powerful systems that dynamically learn and reason at multiple levels of abstraction, from raw data to high-level symbolic reasoning. We will cover use cases and solutions ranging from smart city, transportation, manufacturing, to security and networking.  Back
 
Keywords:
AI Application Deployment and Inference, Advanced AI Learning Techniques (incl. GANs and NTMs), GTC Silicon Valley 2018 - ID S8971
Streaming:
Download:
Share:
 
Accelerating AI Adoption and Impact (Presented by Dell EMC)
Attendees will learn and understand why AI techniques are so powerful, why developing and deploying optimal AI solutions is complex, why using AI techniques effectively is still difficult--and what Dell Technologies is doing to remove these difficult ...Read More
Attendees will learn and understand why AI techniques are so powerful, why developing and deploying optimal AI solutions is complex, why using AI techniques effectively is still difficult--and what Dell Technologies is doing to remove these difficulties and bring easier, effective AI to everyone. Dell Technologies includes seven companies with a comprehensive portfolio of technology products, services and solutions for global industry, government, and education markets, and aims to be the leader in designing and delivering the best AI solutions for every customer, of every type and scale. From Dell Precision workstations for developers and Gateways for edge sensors, to Dell EMC GPU-optimized PowerEdge Servers and Ready Solutions for Deep Learning and hybrid cloud offerings, Dell is leveraging its leadership in technology and in enterprise relationships to design a world-class portfolio of AI solutions for diverse customer workloads, requirements and objectives. This presentation will cover AI and deep learning in an enterprise context, including customer challenges and needs, and then discuss Dell AI solutions and strategy to empower people to use AI rapidly and effectively.  Back
 
Keywords:
AI Application Deployment and Inference, GTC Silicon Valley 2018 - ID S81046
Streaming:
Share:
 
space.ml: Artificial Intelligence Meets Data-Driven Astrophysics
We'll present a suite of artificial intelligence applications and computation geared towards increasing our understanding of the universe. The intensive collaboration between astrophysics and computer science has long started since Jim Gray and Alex ...Read More
We'll present a suite of artificial intelligence applications and computation geared towards increasing our understanding of the universe. The intensive collaboration between astrophysics and computer science has long started since Jim Gray and Alex Szalay. Nowadays, astrophysics continues to offer rich datasets, which are ideal for exploration with the latest in AI and computer science in general. We'll present successful projects in our space.ml initiative that try to answer a range of fascinating astrophysics questions. We'll show how we can use generative adversarial networks to go slightly beyond the Nyquist resolution limit in images, and to study the host galaxies of powerful quasars. We demonstrate how we can use transfer learning to identify rare galaxy mergers, and how to use variational autoencoders to forward model the processes in cosmology and galaxy evolution. We'll illustrate how we can use GPUs for compressive sensing to better analyze data from radio arrays, and to model the evolution of black holes over the age of the universe. Attendees will not only get our current answers to these questions but also get a taste of how AI is reshaping science today.  Back
 
Keywords:
AI Application Deployment and Inference, Astronomy and Astrophysics, GTC Silicon Valley 2018 - ID S8667
Streaming:
Download:
Share:
 
Accelerate TensorFlow Inference with New TensorRT Integration
TensorFlow is an open source software library for numerical computation using data flow graphs. NVIDIA TensorRT is an inference optimizer and runtime for runtime deployment. TensorRT provides optimizations for deep neural networks and uses reduced pr ...Read More
TensorFlow is an open source software library for numerical computation using data flow graphs. NVIDIA TensorRT is an inference optimizer and runtime for runtime deployment. TensorRT provides optimizations for deep neural networks and uses reduced precision to increase throughput, reduce latency, while maintaining accuracy. Today we announced tighter integration in TensorFlow for TensorRT through with new TensorFlow APIs, sub-graph optimizations and INT8 calibration to automatically leverage Tensor Cores on Volta GPUs. TensorRT delivers 2.5x faster inference throughput compared to inference without TensorRT. In this session, NVIDIA developers will use an example based workflow to show how to use this new capability.  Back
 
Keywords:
AI Application Deployment and Inference, Deep Learning and AI Frameworks, GTC Silicon Valley 2018 - ID S81009
Streaming:
Download:
Share:
 
Deep Learning of Railway Track Faults using GPUs
Swiss Federal Railways (SBB) operate a 'diagnosis' train fitted with multiple high-resolution cameras that obtain images of tracks - all while traveling at a speed of 75 mph. Current data processing software conducted in real time on the train prod ...Read More
Swiss Federal Railways (SBB) operate a 'diagnosis' train fitted with multiple high-resolution cameras that obtain images of tracks - all while traveling at a speed of 75 mph. Current data processing software conducted in real time on the train produces  a too high rate of false positives/negatives to the extent that railway experts still need to go on the track to physically inspect anomalies. This is not only very dangerous, but sometimes even impossible and in addition it requires a lot of human labor. We describe how deep learning technologies have been developed to massively improve the automatic detection and classification of railway faults. This is not just a nice-to-have, but rather a must-have in order to ensure the safety of future rail transportation.  Back
 
Keywords:
AI Application Deployment and Inference, Industrial Inspection, GTC Silicon Valley 2018 - ID S8944
Streaming:
Download:
Share:
 
IBM PowerAI: Realizing Business Value with Machine Learning (Presented by IBM)
There is no shortage of hype around AI, but realizing value through machine and deep learning comes with its challenges. IBM PowerAI removes the inhibitors across each stage of a workflow, allowing enterprises to rapidly realize business value with A ...Read More
There is no shortage of hype around AI, but realizing value through machine and deep learning comes with its challenges. IBM PowerAI removes the inhibitors across each stage of a workflow, allowing enterprises to rapidly realize business value with AI.  Back
 
Keywords:
AI Application Deployment and Inference, GTC Silicon Valley 2018 - ID S81048
Streaming:
Share:
 
NVIDIA GPU Video Technologies and Video Codec SDK: Updates and Roadmap
NVIDIA's video SDK is a set of APIs for hardware-accelerated video encoding and decoding using NVIDIA GPUs. We'll provide an overview of the APIs, with particular emphasis on the latest features, such as FFmpeg support of NVIDIA-accelerated transco ...Read More
NVIDIA's video SDK is a set of APIs for hardware-accelerated video encoding and decoding using NVIDIA GPUs. We'll provide an overview of the APIs, with particular emphasis on the latest features, such as FFmpeg support of NVIDIA-accelerated transcoding, quality and performance enhancements. We'll discuss some strategies on efficient usage of GPU video hardware acceleration for use cases such as video inferencing, transcoding, and media archiving.  Back
 
Keywords:
AI Application Deployment and Inference, Video and Image Processing, GTC Silicon Valley 2018 - ID S8601
Streaming:
Download:
Share:
 
Monitoring Honey Bee Health Using TensorRT and Microsoft Cognitive Toolkit
We'll take a deep dive into honey bee hive health monitoring with NVIDIA's TX2, TensorRT (a high-performance deep learning inference optimizer), Kineticas insight engine running on DGX-1/DGXStaion, and Microsoft Cognitive Toolkit to rapidly o ...Read More
We'll take a deep dive into honey bee hive health monitoring with NVIDIA's TX2, TensorRT (a high-performance deep learning inference optimizer), Kineticas insight engine running on DGX-1/DGXStaion, and Microsoft Cognitive Toolkit to rapidly optimize, validate, and deploy trained neural networks for inference. In recent years, the media has reported that bees seem to be dying at an unprecedented rate. We'll explore how new accelerated analytics technologies and their corresponding compute platforms can deliver game-changing possibilities for innovation as we follow a honey bee farm scientist in California, who agreed to field test this real-time monitoring solution with her beehives.  See first-hand how adaptable and accessible these complex, cutting-edge technologies have become and how we can use intelligent monitoring technologies to help rescue the honey bee in the real-world environmental analytics opportunity.  Back
 
Keywords:
AI Application Deployment and Inference, Deep Learning and AI Frameworks, GTC Silicon Valley 2018 - ID S8508
Streaming:
Share:
 
Practical Application of Deep Learning in Smart Factory: Visual Inspection System of Semiconductor Laser
Fujikura is pushing forward of implementation of the smart factory with AI and IoT for improving the productivity and production quality. In this presentation, we will present visual inspection system incorporating deep learning in the production pro ...Read More
Fujikura is pushing forward of implementation of the smart factory with AI and IoT for improving the productivity and production quality. In this presentation, we will present visual inspection system incorporating deep learning in the production process of semiconductor lasers. Not only OK/NG classification, but also multiple NG mode classification was performed. The inspection accuracy of 95 % that is equivalent to skilled workers' accuracy was achieved by optimizing the data set and the hyper parameters of a CNN model. The activation map was used for reliability and validity assurance. We will present the difficulty in our practical application in manufacturing industry, such as the small number of some category and small defect/chip size ratio, and also introduce our countermeasures.  Back
 
Keywords:
AI Application Deployment and Inference, Industrial Inspection, GTC Silicon Valley 2018 - ID S8911
Streaming:
Download:
Share:
 
Deep Learning for Heliophysics
NASA's heliophysics division operates a fleet of spacecraft, the so-called Heliophysics System Observatory, to monitor the Sun's activity and how its changes drive space weather in interplanetary space and in the near-Earth environment. We'll pres ...Read More
NASA's heliophysics division operates a fleet of spacecraft, the so-called Heliophysics System Observatory, to monitor the Sun's activity and how its changes drive space weather in interplanetary space and in the near-Earth environment. We'll present case studies of how a number of challenging problems encountered in heliophysics can be tackled using deep learning: spectropolarimetric inversions for measuring the magnetic field on the solar surface, and mega-Kelvin thermometry of the Sun's corona by using a deep neural network to solve a compressed sensing problem. These low-cost solutions make possible new concepts for deep space missions for space weather monitoring. Some of the work in this presentation was made possible by NASA's Frontier Development Lab, a public-private partnership between the agency and industry partners (including the SETI Institute, NVIDIA, IBM, Intel, kx & Lockheed Martin), whose mission is to use artificial intelligence to tackle problems related to planetary defense and heliophysics.  Back
 
Keywords:
AI Application Deployment and Inference, Accelerated Analytics, Astronomy and Astrophysics, GTC Silicon Valley 2018 - ID S8222
Streaming:
Download:
Share:
 
Distributed and Scalable Video Analytics on Tegra X1/X2 Based Embedded Computer Cluster
A wide area and city surveillance system solution for running real-time video analytics on thousands of 1080p video streams will be presented. System hardware is an embedded computer cluster based on NVIDIA TX1/TX2 and NXP iMX6 modules. A custom ...Read More

A wide area and city surveillance system solution for running real-time video analytics on thousands of 1080p video streams will be presented. System hardware is an embedded computer cluster based on NVIDIA TX1/TX2 and NXP iMX6 modules. A custom designed system software manages job distribution, resulting in collection and system wide diagnostics including instantaneous voltage, power and temperature readings. System is fully integrated with a custom designed video management software, IP cameras and network video recorders. Instead of drawing algorithm results on the processed video frames, re-encoding and streaming back to the operator computer for display, only the obtained metadata is sent to the operator computer. Video management software streams video sources independently, and synchronizes decoded video frames with the corresponding metadata locally, before presenting the processed frames to the operator.

  Back
 
Keywords:
AI Application Deployment and Inference, Intelligent Video Analytics and Smart Cities, GTC Silicon Valley 2018 - ID S8409
Streaming:
Download:
Share:
 
How AI Technology Lifts the Ads Business in JD.com
Deep learning and reinforcement learning are widely used in ads products of JD.com, e.g. ranking model in recommender systems, bidding model in ad exchange business and automatic ads review systems. These technologies have brought great benefits to J ...Read More
Deep learning and reinforcement learning are widely used in ads products of JD.com, e.g. ranking model in recommender systems, bidding model in ad exchange business and automatic ads review systems. These technologies have brought great benefits to JD.com and all of them are built on Nvidia GPUs.  Back
 
Keywords:
AI Application Deployment and Inference, Consumer Engagement and Personalization, GTC Silicon Valley 2018 - ID S81016
Streaming:
Download:
Share:
 
A Map of Knowledge: Using Behavioral Data in Higher-Ed to Surface Novel Semantic Structure and Personalized Guidance
Personalized learning has been a promising but often elusive ideal sought after in education. We'll demonstrate the progress made with two concrete examples of personalized learning supports implemented at scale in a massive open online course (MOOC ...Read More
Personalized learning has been a promising but often elusive ideal sought after in education. We'll demonstrate the progress made with two concrete examples of personalized learning supports implemented at scale in a massive open online course (MOOC) and on the UC Berkeley campus in a collaboration with the Office of the Registrar. Both approaches employ long short-term memory to leverage a collaborative signal out of millions of historic learner actions. In the case of the MOOC, the next page a learner is expected to spend considerable time on is predicted and offered as a real-time suggestion. At the university, we consider sequences of millions of historic enrollments over the past eight years. These sequences of course identifiers, when modeled with representation learning approaches most commonly applied to natural language, reveal a tremendous degree of semantic relational information about the courses which can be visualized, reasoned about, and surfaced to students. Our course information platform uses this automatically inferred semantic information to help students navigate the university's offerings and provides personalized course suggestions based on topic preference.  Back
 
Keywords:
AI Application Deployment and Inference, Consumer Engagement and Personalization, AI and DL Research, GTC Silicon Valley 2018 - ID S8597
Streaming:
Share:
 
Pioneering AI for All
Businesses of all sizes are increasingly recognizing the potential value of AI, but few are sure how to prepare for the transformational change it is sure to bring to their organizations. Danny Lange rolled out company-wide AI platforms at Uber ...Read More

Businesses of all sizes are increasingly recognizing the potential value of AI, but few are sure how to prepare for the transformational change it is sure to bring to their organizations. Danny Lange rolled out company-wide AI platforms at Uber and Amazon; now, through Unity Technologies, he's making AI available to the rest of us. He'll also share his thoughts for the most exciting advances that AI will bring over the next year. His insights will help you understand the true potential of AI, regardless of your role or industry.

  Back
 
Keywords:
AI Application Deployment and Inference, Advanced AI Learning Techniques (incl. GANs and NTMs), AI and DL Business Track (high level), AI for Business, GTC Silicon Valley 2018 - ID S8729
Streaming:
Share:
 
Low-Latency GPU Accelerated Inferencing with TensorRT
Come learn how you can optimize the deployment of your trained neural networks using the GPU-accelerated inferencing library called TensorRT. TensorRT is a high-performance tool for low-latency, high-throughput deep neural network (DNN) inference tha ...Read More
Come learn how you can optimize the deployment of your trained neural networks using the GPU-accelerated inferencing library called TensorRT. TensorRT is a high-performance tool for low-latency, high-throughput deep neural network (DNN) inference that runs on NVIDIA GPUs. The latest release of TensorRT introduces a novel, framework-agnostic network definition format called universal framework format, allowing TensorRT to support and optimize DNN models trained in multiple deep learning frameworks like Caffe and TensorFlow. It also provides the capability to run inference at reduced precision, giving developers the ability to take advantage of new GPU hardware features like the Volta Tensor Core architecture. This session will be a combination of lecture and live demos.  Back
 
Keywords:
AI Application Deployment and Inference, Tools and Libraries, Performance Optimization, Data Center and Cloud Infrastructure, GTC Silicon Valley 2018 - ID S8496
Streaming:
Share:
 
Intelligent Talent Management - AI Drives Transformation
Artificial intelligence helps you hire faster and smarter. It also helps you determine your career path, learning, and development. Wondering how? AI platforms have a brain that reads, understands, and analyzes just as human beings do. They can read ...Read More
Artificial intelligence helps you hire faster and smarter. It also helps you determine your career path, learning, and development. Wondering how? AI platforms have a brain that reads, understands, and analyzes just as human beings do. They can read thousands and millions of resumes, JDs, career progressions, and learning content in a matter of seconds. This equips them with intelligence creating a neural network of skills, demographics, industries, occupations, and courses/certifications. This acts as the central intelligence powering search and match algorithms to find accurate matches to job demands in a few seconds. The NLP layer helps understand intent, for example, it differentiates between 'Worked with a PM' and 'Worked as a PM' to determine that the former could work collaboratively and the latter could drive projects. AI platforms mimic a recruiter or hiring manager's brain to find that right match. What takes HR 20-30 days is done in a few seconds by an AI platform. It helps HR leaders in workforce planning by forecasting what skills and domains to invest, maintain, or upgrade in their organizations, which could be a game changer especially for people-centric organizations.  Back
 
Keywords:
AI Application Deployment and Inference, Accelerated Analytics, AI and DL Research, AI and DL Business Track (high level), GTC Silicon Valley 2018 - ID S8303
Streaming:
Share:
 
Deploying, Profiling, and Optimizing Distributed TensorFlow in Production with GPUs
Using the latest advancements from TensorFlow including the Accelerated Linear Algebra (XLA) Framework, JIT/AOT Compiler, and Graph Transform Tool, we'll demonstrate how to optimize, profile, and deploy TensorFlow models in GPU-based production envi ...Read More
Using the latest advancements from TensorFlow including the Accelerated Linear Algebra (XLA) Framework, JIT/AOT Compiler, and Graph Transform Tool, we'll demonstrate how to optimize, profile, and deploy TensorFlow models in GPU-based production environments. We'll cover many demos based on open source tools. You can completely reproduce all demos through Docker on your own GPU cluster. See http://pipeline.ai for links to the GitHub Repo.  Back
 
Keywords:
AI Application Deployment and Inference, NVIDIA Inception Program, Deep Learning and AI Frameworks, GTC Silicon Valley 2018 - ID S8621
Streaming:
Share:
 
Latest Tools and Techniques for Training and Deploying Deep Neural Networks in Educational Environments
Craig Morioka, UCLA Adjunct Associate Professor of Radiological Sciences, and Dima Lituiev, Postdoctoral Scholar at the University of California San Francisco, Institute for Computational Health Sciences, will discuss how they empower their fellow fa ...Read More
Craig Morioka, UCLA Adjunct Associate Professor of Radiological Sciences, and Dima Lituiev, Postdoctoral Scholar at the University of California San Francisco, Institute for Computational Health Sciences, will discuss how they empower their fellow faculty, staff, and students with the latest techniques in training and deploying deep neural networks through NVIDIAs Deep Learning Institute (DLI) University Ambassador Program - a new AI and Deep Learning education enablement program for universities. This will include a dive into the benefits of an online learning platform, which uses GPUs in the cloud, by stepping through the DLIs online Image Segmentation and Radiomics labs. The Image Segmentation lab leverages an example from medical image analysis where it is often important to separate pixels corresponding to different types of tissue or cells for the purposes of diagnostics and treatment planning. Dima uses image segmentation in his research to facilitate diagnostics of kidney rejection by analyzing histological slides from patients with kidney transplants. We will explore how the Tensorflow code is structured and how the Tensorboard tool can be used to visualize structure and training dynamics of segmentation models. The focus of the Radiomics lab is detection of the 1p19q co-deletion biomarker using deep learning - specifically convolutional neural networks using the Keras and TensorFlow computing frameworks. Attendees will also learn how they can apply to become a DLI University Ambassador and bring the latest in Deep Learning and AI education to their academic communities.    Back
 
Keywords:
AI Application Deployment and Inference, Deep Learning and AI Frameworks, AI and DL Business Track (high level), GTC Silicon Valley 2018 - ID S8823
Streaming:
Share:
 
Protecting Pulsed High-Power Lasers with Real-Time Image Classification
Learn how to combine computer vision techniques and deep learning to improve the sensitivity of a real-time, GPU-powered safety system. In petawatt laser systems, firing at 10 Hz, suddenly appearing scatterers can damage components. Spreading of dama ...Read More
Learn how to combine computer vision techniques and deep learning to improve the sensitivity of a real-time, GPU-powered safety system. In petawatt laser systems, firing at 10 Hz, suddenly appearing scatterers can damage components. Spreading of damage can be avoided by suspending operation immediately on occurrence of such an event. We'll present our approach for the automatic detection of critical failure states from intensity profiles of the laser beam. By incorporating quick feature detection and learned heuristics for feature classification, both real-time constraints and limited available training data are accommodated. Localization of triggering feature is crucial for when the problem is located in non-sensitive sections and will not be removed from the beam in production.  Back
 
Keywords:
AI Application Deployment and Inference, Advanced AI Learning Techniques (incl. GANs and NTMs), Computer Vision, GTC Silicon Valley 2018 - ID S8330
Streaming:
Share:
 
Driver Drowsiness Detection for ADAS
We'll present an in-car ADAS technology to detect drowsy driving. This technique can be used to alert and awaken the driver, or take corrective actions if required. We employ a CNN-based approach for this technique, which is trained on a mix of synt ...Read More
We'll present an in-car ADAS technology to detect drowsy driving. This technique can be used to alert and awaken the driver, or take corrective actions if required. We employ a CNN-based approach for this technique, which is trained on a mix of synthetic and real images. We'll cover the details of the detection system pipeline and the synthetic dataset generation. We'll also show a demonstration of the detection system in action.  Back
 
Keywords:
AI Application Deployment and Inference, Autonomous Vehicles, GTC Silicon Valley 2018 - ID S8399
Streaming:
Share:
 
Deep Learning Demystified
What is Deep Learning? In what fields is it useful? How does it relate to artificial intelligence? We'll discuss  deep learning and why this powerful new technology is getting so much attention, learn how deep neural networks are traine ...Read More

What is Deep Learning? In what fields is it useful? How does it relate to artificial intelligence? We'll discuss  deep learning and why this powerful new technology is getting so much attention, learn how deep neural networks are trained to perform tasks with super-human accuracy, and the challenges organizations face in adopting this new approach. We'll also cover some of the best practices, software, hardware, and training resources that many organizations are using to overcome these challenges and deliver breakthrough results.

  Back
 
Keywords:
AI Application Deployment and Inference, Deep Learning and AI Frameworks, Deep Learning and AI, GTC Silicon Valley 2018 - ID S8669
Streaming:
Share:
 
CatBoost: Fast Open-Source Gradient Boosting Library For GPU
Learn how to use GPUs to accelerate gradient boosting on decision trees. We'll discuss CUDA implementation of CatBoost an open-source library that successfully handles categorical features and shows better quality compared to other open-source gra ...Read More
Learn how to use GPUs to accelerate gradient boosting on decision trees. We'll discuss CUDA implementation of CatBoost an open-source library that successfully handles categorical features and shows better quality compared to other open-source gradient boosted decision trees libraries. We'll provide a brief overview of problems which could be solved with CatBoost. Then, we'll discuss challenges and key optimizations in the most significant computation blocks. We'll describe how one can efficiently build histograms in shared memory to construct decision trees and how to avoid atomic operation during this step. We'll provide benchmarks that shows that our GPU implementation is five to 40 times faster compared to Intel server CPUs. We'll also provide performance comparison against GPU implementations of gradient boosting in other open-source libraries.  Back
 
Keywords:
AI Application Deployment and Inference, Tools and Libraries, HPC and AI, GTC Silicon Valley 2018 - ID S8393
Streaming:
Share:
 
Leveraging GPUs for Bayesian Inference
We'll present results on speeding up Bayesian inference in NVIDIA DGX-1 server for medical diagnostics. Bayesian inference is an AI technique to reason under uncertainty that is computationally and data intensive. We'll discuss the implications for ...Read More
We'll present results on speeding up Bayesian inference in NVIDIA DGX-1 server for medical diagnostics. Bayesian inference is an AI technique to reason under uncertainty that is computationally and data intensive. We'll discuss the implications for both inference and training of Bayesian networks.  Back
 
Keywords:
AI Application Deployment and Inference, Accelerated Analytics, GTC Silicon Valley 2018 - ID S8488
Streaming:
Share:
 
Prototyping Vision-Based Classifiers in Constrained Environments
SOFWERX developed a vision-based classifier using commodity hardware and machine learning libraries to satisfy an urgent high-level requirement. To track the usage of tank ammunition, the team had to address challenges involving unavailable training ...Read More
SOFWERX developed a vision-based classifier using commodity hardware and machine learning libraries to satisfy an urgent high-level requirement. To track the usage of tank ammunition, the team had to address challenges involving unavailable training data, varying spatial orientations, and limited power consumption. To resolve these challenges, SOFWERX generated an augmented dataset using synthetic models, implemented spatial transformers, and experimented with different hardware/software optimizations.  Back
 
Keywords:
AI Application Deployment and Inference, Performance Optimization, GTC Silicon Valley 2018 - ID S8193
Streaming:
Download:
Share:
 
Enabling Deep Learning Applications in Radio Frequency Systems
Artificial intelligence has made great strides in many technology sectors, however, it has yet to impact the design and applications of radio frequency (RF) and wireless systems. This is primarily due to the industry''s preference towards field progr ...Read More
Artificial intelligence has made great strides in many technology sectors, however, it has yet to impact the design and applications of radio frequency (RF) and wireless systems. This is primarily due to the industry''s preference towards field programmable gate array (FPGA) systems. Conversely, the deep learning revolution has been fueled by GPUs and the ease with which they may be programmed for highly parallel computations. The next generation RF and wireless technology will require wide-band systems capable of real-time machine learning with GPUs. Working with strategic partners, we''ve designed a software configurable wide-band RF transceiver system capable of performing real-time signal processing and machine learning with a Jetson TX2. We discuss system performance, collection of RF training data, and the software used by the community to create custom applications. Additionally, we''ll present data demonstrating applications in the field of RF machine learning and deep learning.  Back
 
Keywords:
AI Application Deployment and Inference, NVIDIA Inception Program, Cyber Security, IoT, Robotics & Autonomous Machines, GTC Silicon Valley 2018 - ID S8375
Streaming:
Share:
 
Performance Optimization for Deep Image Matting in Photoshop
Learn how a research paper from Adobe Research Labs makes it into a real customer product like Photoshop. We attempted to solve a number of challenging issues about applying the technology to real-world use cases, including large model size, heavy me ...Read More
Learn how a research paper from Adobe Research Labs makes it into a real customer product like Photoshop. We attempted to solve a number of challenging issues about applying the technology to real-world use cases, including large model size, heavy memory consumption, and slow runtime performance.  Back
 
Keywords:
AI Application Deployment and Inference, GTC Silicon Valley 2018 - ID S8550
Streaming:
Download:
Share:
 
Optimizing NMT with TensorRT
OpenNMT is an open source neural machine translation and neural machine sequencing model. Using Volta Tensor Cores and TensorRT, we''re able to improve performance by 100 times over CPU implementation. We''ll discuss OpenNMT and how we implement it v ...Read More
OpenNMT is an open source neural machine translation and neural machine sequencing model. Using Volta Tensor Cores and TensorRT, we''re able to improve performance by 100 times over CPU implementation. We''ll discuss OpenNMT and how we implement it via TensorRT. We''ll show how by using our plugin interface and new TensorRT features, we''re able to implement this network at high performance.  Back
 
Keywords:
AI Application Deployment and Inference, Advanced AI Learning Techniques (incl. GANs and NTMs), GTC Silicon Valley 2018 - ID S8822
Streaming:
Download:
Share:
 
Breaking the Barriers to AI-Scale in the Enterprise
Organizations everywhere want to AI-infuse every aspect of their business, but need a platform that delivers the scale and flexibility to fit both IT operational constraints, as well as workload performance demanded by data scientists. Attend this se ...Read More
Organizations everywhere want to AI-infuse every aspect of their business, but need a platform that delivers the scale and flexibility to fit both IT operational constraints, as well as workload performance demanded by data scientists. Attend this session to get see the latest advancements in scaling in GPU servers and deep learning software, and hear how the latest solutions from NVIDIA solve your biggest AI platform challenges  Back
 
Keywords:
AI Application Deployment and Inference, Data Center and Cloud Infrastructure, AI and DL Research, GTC Silicon Valley 2018 - ID S8196
Streaming:
Download:
Share:
 
Continuous Delivery of AI Applications
Deep learning systems are usually developed by data scientists, who are good at mathematics and computer science. But to deploy and operationalize these models for broader use, you need the devops mindset and tools. We''ll show you how to connect the ...Read More
Deep learning systems are usually developed by data scientists, who are good at mathematics and computer science. But to deploy and operationalize these models for broader use, you need the devops mindset and tools. We''ll show you how to connect the workflow between the data scientists and devops. We''ll also explore basic continuous integration and delivery concepts and how they can be applied to deep learning models. Using a number of AWS services, we''ll showcase how you can take the output of a deep learning model and deploy it to perform predictions in real time with low latency and high availability. In particular, we''ll showcase the ease of deploying DL to predict functions using Apache MXNet (a deep learning library), Amazon ECS, Amazon S3, and Amazon ECR, Amazon developer tools, and AWS CloudFormation.  Back
 
Keywords:
AI Application Deployment and Inference, GTC Silicon Valley 2018 - ID S8173
Streaming:
Download:
Share:
 
Defect Inspection from Scratch to Production
In order to fulfill customer''s requirement, companies have to guarantee the quality of delivered products, which can often be achieved only by manually inspection of the finished product. Since human-based defect inspection and classification are ti ...Read More
In order to fulfill customer''s requirement, companies have to guarantee the quality of delivered products, which can often be achieved only by manually inspection of the finished product. Since human-based defect inspection and classification are time-consuming and the results vary by individuals, automatic defect detection and classification has the potential to reduce the cost of quality assurance significantly. In this talk, we will demonstrate how to utilize deep learning algorithms, i.e., Fully Convolutional Neural Network to build a general defect inspection and classification model. We will also share experiences on how to effectively collect labelling data, deal with imbalance data, and also how to optimize the model in terms of latency and throughput with TensorRT before deploy the model to the production line.  Back
 
Keywords:
AI Application Deployment and Inference, Industrial Inspection, IoT, Robotics & Drones, Robotics & Autonomous Machines, GTC Silicon Valley 2018 - ID S8682
Streaming:
Download:
Share:
 
Identifying Defect Patterns in Hard Disk Drive Magnetic Media Manufacturing Processes Using Real and Synthetic Data
Learn how synthetic data can be used to develop traditional and Convolutional Neural Network (CNN) image segmentation models when labelled training data is limited. We will describe hard drive media defect patterns and how they relate to problems i ...Read More
Learn how synthetic data can be used to develop traditional and Convolutional Neural Network (CNN) image segmentation models when labelled training data is limited. We will describe hard drive media defect patterns and how they relate to problems in the manufacturing line, show why CNN models were chosen for some defect patterns, and how the CNN models were trained using both synthetic and real data. Different architectures using CNNs were explored and the resulting benefits and drawbacks are presented.  Back
 
Keywords:
AI Application Deployment and Inference, Industrial Inspection, IoT, Robotics & Drones, Robotics & Autonomous Machines, GTC Silicon Valley 2018 - ID S8415
Streaming:
Download:
Share:
 
Using AI for Interactive Applications
Machine learning has revolutionized many important fields, ranging from computer vision and natural language processing to healthcare and robotics. In this session, we will discuss how developers can embrace machine learning methods for graphics and ...Read More
Machine learning has revolutionized many important fields, ranging from computer vision and natural language processing to healthcare and robotics. In this session, we will discuss how developers can embrace machine learning methods for graphics and gaming. We''ll cover both gaming use cases and general applications of machine learning as well as how to best leverage recent GPU hardware for machine learning workloads.  Back
 
Keywords:
AI Application Deployment and Inference, Graphics and AI, AI for Gaming, Rendering and Ray Tracing, GTC Silicon Valley 2018 - ID S8957
Streaming:
Download:
Share:
 
Anomaly Detection on Vehicle CAN BUS
We''ll discuss anomaly detection on vehicle CAN BUS. We developed a novel solution for neural networks to detect anomalies in CAN data. Due to the inherent characteristics of controller area (CAN) networks, such as lack of authentication and followin ...Read More
We''ll discuss anomaly detection on vehicle CAN BUS. We developed a novel solution for neural networks to detect anomalies in CAN data. Due to the inherent characteristics of controller area (CAN) networks, such as lack of authentication and following a broadcast routing scheme, devices connected to a CAN network are exposed to a broad range of cyberattacks. Our work aims to alleviate this problem by providing an anomaly detection mechanism, that is, identifying deviations from normal network traffic, to enhance the security of CAN networks. This invention is leveraged as one of the intrusion detection methods in a broader NVIDIA embedded software security system deployed in automotive platforms. In this specific application, the embedded system is a car computer -- an embedded system deployed in modern vehicles. Typical examples: infotainment systems, ADAS units, dashboards, head units. The vulnerable endpoints are all the peripherals connected to the computer. Typical examples: sensors, cameras, media devices, local and wide area communication interfaces and devices (for example, WiFi, BT, cellular), specific car network interfaces and devices.  Back
 
Keywords:
AI Application Deployment and Inference, Deep Learning and AI Frameworks, Cyber Security, Autonomous Vehicles, GTC Silicon Valley 2018 - ID S8347
Streaming:
Download:
Share:
 
Highly-Efficient Caching with Tiling & Chaining in CNN
Learn how to achieve 100% R/W cache hit rate for most intermediate tensors in CNN and over 80% typical DRAM traffic saving, with general applicability to a limited cache size and large tensors. The high-throughput NVIDIA Tensor Core and DLA demand hi ...Read More
Learn how to achieve 100% R/W cache hit rate for most intermediate tensors in CNN and over 80% typical DRAM traffic saving, with general applicability to a limited cache size and large tensors. The high-throughput NVIDIA Tensor Core and DLA demand high memory traffic. Chaining of consecutive layers in CNN can save DRAM traffic by reusing intermediate tensors in cache. This strategy is effective only with small tensors and a large cache. In this work, we slice tensors into small tiles (with halo) and chain these tiles so the requirement for perfect caching can always be fulfilled. Our implementation of this approach is proven to be very effective in saving DRAM traffic. This work allows us to solve the memory bandwidth issue of CNN with a relatively small but high-bandwidth cache.  Back
 
Keywords:
AI Application Deployment and Inference, Performance Optimization, GTC Silicon Valley 2018 - ID S8299
Streaming:
Share:
 
Scalable, Responsive, and Cost-Effective Object Detection Service for Web-Scale Images
We''ll introduce how Bing built a scalable, responsive, and economical object detection API based on NVIDIA GPUs and Azure cloud platforms. Object detection is an important image understanding technique as the entry point or dispatcher to guide users ...Read More
We''ll introduce how Bing built a scalable, responsive, and economical object detection API based on NVIDIA GPUs and Azure cloud platforms. Object detection is an important image understanding technique as the entry point or dispatcher to guide users to more specific scenarios. However, it is very challenging to provide object detection services on web-scale images because it is intrinsically a compute-intensive task and thus resource demanding. We''ll also introduce how to use NVIDIA''s CUDA profiling toolchain and cuDNN to make the system even more cost-effective. The system currently supports billion-level traffic, covering Bing''s entire index.  Back
 
Keywords:
AI Application Deployment and Inference, Performance Optimization, GTC Silicon Valley 2018 - ID S8620
Streaming:
Download:
Share:
 
Revisiting the TurboCharged Test Toolbox: VR, Robotics, and More DL
Last year, we began to see promising results of applying Deep Learning in an unexpected space: hardware QA. Fast forward +365, and the efforts have been to expand on what we''ve learned, push the technology broader and into other areas that will ulti ...Read More
Last year, we began to see promising results of applying Deep Learning in an unexpected space: hardware QA. Fast forward +365, and the efforts have been to expand on what we''ve learned, push the technology broader and into other areas that will ultimately aid in our greatest challenge: testing at scale. In this session we will highlight a new piece of the problem we are tackling: VR. We will introduce methodologies for not only addressing the unique problems that VR testing presents, but will also showcase some of the other test process areas where we are applying other Deep Learning models to gain efficiency in our overall production pipeline. From using DL on our bug mining to create a quicker path from tester to developer and back, to analysis on end user issues as a method for task automation, explore how we are enabling speed, accuracy and efficiency.  Back
 
Keywords:
AI Application Deployment and Inference, Virtual Reality and Augmented Reality, Tools and Libraries, Graphics and AI, AI for Gaming, GTC Silicon Valley 2018 - ID S8262
Streaming:
Share:
 
Building Seeing AI : The Talking Camera App for the Blind
We''ll detail the journey of building Seeing AI, an app from Microsoft AI & Research that narrates the world around you. Designed for the blind and low-vision community, this research project harnesses the power of AI to describe people, text, an ...Read More
We''ll detail the journey of building Seeing AI, an app from Microsoft AI & Research that narrates the world around you. Designed for the blind and low-vision community, this research project harnesses the power of AI to describe people, text, and objects. Seeing AI leverages object classification, detection, image captioning, and more, with several running on the device in real time at more than 15 frames per second. We''ll go over the learnings, challenges, hits, and misses we encountered while developing the application.  Back
 
Keywords:
AI Application Deployment and Inference, Computer Vision, GTC Silicon Valley 2018 - ID S8598
Streaming:
Share:
 
Deep Learning Infrastructure for Autonomous Vehicles
We''ll introduce deep learning infrastructure for building and maintaining autonomous vehicles, including techniques for managing the lifecycle of deep learning models, from definition, training and deployment to reloading and life-long ...Read More

We''ll introduce deep learning infrastructure for building and maintaining autonomous vehicles, including techniques for managing the lifecycle of deep learning models, from definition, training and deployment to reloading and life-long learning. DNN autocurates and pre-labels data in the loop. Given data, it finds the best run-time optimized deep learning models. Training scales with data size beyond multi-nodes. With these methodologies, one takes only data from the application and feeds DL predictors to it. This infrastructure is divided into multiple tiers and is modular, with each of the modules containerized to lower infrastructures like GPU-based cloud infrastructure.

  Back
 
Keywords:
AI Application Deployment and Inference, Data Center and Cloud Infrastructure, Autonomous Vehicles, Autonomous Machines, GTC Silicon Valley 2018 - ID S8531
Streaming:
Download:
Share:
 
Deploying Machine Learning on the Oilfield: From the Labs to the Edge
Deploying machine learning-based predictive models to the oil field is quite challenging. They are remote, hazardous, and have spotty connectivity to the cloud. The world of operationalizing a model is very different than the perfect lab environment ...Read More
Deploying machine learning-based predictive models to the oil field is quite challenging. They are remote, hazardous, and have spotty connectivity to the cloud. The world of operationalizing a model is very different than the perfect lab environment where the models are born. We'll detail the requirements of our oil and gas customers and how we were able to meet those requirements such that we could deploy a new generation of analytics with a complete software engineering discipline and mentality around it by taking advantage of the Microsoft IoT Edge platform. This is currently a pilot project under way and, due to the engineering principals in place, we are able to complete a loop from the field to the lab and back again.  Back
 
Keywords:
AI Application Deployment and Inference, IoT, Robotics & Drones, Robotics & Autonomous Machines, GTC Silicon Valley 2018 - ID S8714
Streaming:
Download:
Share:
 
Digital Twin for the Railway Network
We describes concept of Digital Twin with respect to the Railway Network. Railroad customers across the world manage thousands of miles of Track infrastructure that consists of the Rails, Ballast, Ties, Bridges, Tunnels, Wayside equipment, etc. This ...Read More
We describes concept of Digital Twin with respect to the Railway Network. Railroad customers across the world manage thousands of miles of Track infrastructure that consists of the Rails, Ballast, Ties, Bridges, Tunnels, Wayside equipment, etc. This talk demonstrates a new approach to Track infrastructure monitoring that GE is piloting for customers using the concept of Digital Twin for network. Using an offline GPU infrastructure Deep Learning models are created and trained on large volumes of video data to learn the state of healthy Track and predict anomalies. During the talk, real customer use-case videos will be shown that demonstrate Analytics on videos from Locomotive-mounted cameras with Deep Learning models to calculate health index and display on a map for driving Maintenance decisions.  Back
 
Keywords:
AI Application Deployment and Inference, Computer Vision, GTC Silicon Valley 2018 - ID S8614
Streaming:
Download:
Share:
 
How Deep Learning Could Predict Weather Events
How do meteorologists predict weather or weather events such as hurricanes, typhoons, and heavy rain? Predicting weather events were done based on supercomputer (HPC) simulations using numerical models such as WRF, UM, and MPAS. But recently, many de ...Read More
How do meteorologists predict weather or weather events such as hurricanes, typhoons, and heavy rain? Predicting weather events were done based on supercomputer (HPC) simulations using numerical models such as WRF, UM, and MPAS. But recently, many deep learning-based researches have been showing various kinds of outstanding results. We'll introduce several case studies related to meteorological researches. We'll also describe how the meteorological tasks are different from general deep learning tasks, their detailed approaches, and their input data such as weather radar images and satellite images. We'll also cover typhoon detection and tracking, rainfall amount prediction, forecasting future cloud figure, and more.  Back
 
Keywords:
AI Application Deployment and Inference, Climate, Weather, Ocean Modeling, Computer Vision, HPC and AI, GTC Silicon Valley 2018 - ID S8816
Streaming:
Download:
Share:
 
Visual Search at eBay
We'll share information and lessons learned from developing a scalable visual search engine to handle a massive volatile inventory like eBay. We'll describe how eBay data is challenging for visual search, how to leverage a single deep neural networ ...Read More
We'll share information and lessons learned from developing a scalable visual search engine to handle a massive volatile inventory like eBay. We'll describe how eBay data is challenging for visual search, how to leverage a single deep neural network to perform multiple tasks efficiently, how to deploy our solution in a distributed cloud infrastructure, and which optimizations we have made for a trade-off between relevance and latency. We'll give examples and insights to benefit computer vision practitioners in the industry who intend to build up visual search engines from scratch.  Back
 
Keywords:
AI Application Deployment and Inference, Data Center and Cloud Infrastructure, Computer Vision, GTC Silicon Valley 2018 - ID S8766
Streaming:
Share:
 
The Long Road to Model Deployment or how to make a good model great!
In this talk we will cover the essential building blocks of the AI platform Nvidia engineers are using to build a world-class automotive perception stack. Through a computer vision application example, we will see how to improve a baseline model to p ...Read More
In this talk we will cover the essential building blocks of the AI platform Nvidia engineers are using to build a world-class automotive perception stack. Through a computer vision application example, we will see how to improve a baseline model to produce better, faster predictions. The talk will focus on: - hyper-parameter optimization, - model complexity reduction (pruning), - target platform optimizations (TensorRT integration), - automation of complex workflows  Back
 
Keywords:
AI Application Deployment and Inference, Deep Learning and AI Frameworks, GTC Silicon Valley 2018 - ID S8633
Streaming:
Download:
Share:
 
Containerizing Deep Learning with Singularity
We'll talk about how to use Singularity to containerize deep learning applications. We'll provide compelling reasons to choose Singularity over Docker. We'll cover deep learning frameworks, including TensorFlow, NV-Caffe, MXNet, and others. We'll ...Read More
We'll talk about how to use Singularity to containerize deep learning applications. We'll provide compelling reasons to choose Singularity over Docker. We'll cover deep learning frameworks, including TensorFlow, NV-Caffe, MXNet, and others. We'll present the current challenges and workarounds when using Singularity in a HPC cluster. We'll compare the performance of Singularity to bare-metal systems.  Back
 
Keywords:
AI Application Deployment and Inference, HPC and AI, GTC Silicon Valley 2018 - ID S8368
Streaming:
Download:
Share:
 
ANI-AL: Universal Deep Learning Potentials for Organic Molecules and Materials
We'll introduce ANI-AL molecular potentials, which are deep learning based potential energy functions for the fast and accurate prediction of quantum mechanical energies and forces of molecular systems. Thanks to GPU acceleration of training and inf ...Read More
We'll introduce ANI-AL molecular potentials, which are deep learning based potential energy functions for the fast and accurate prediction of quantum mechanical energies and forces of molecular systems. Thanks to GPU acceleration of training and inference, we successfully implement an automated sampling method that borrows techniques from active learning to automatically drive the systematic improvement of ANI-AL potentials. We'll also present results from applications of the ANI-AL potential in various problems relating to computational chemistry, such as molecular structure optimization, reaction path prediction, vibrational frequency calculation, and molecular dynamics simulations.  Back
 
Keywords:
AI Application Deployment and Inference, Computational Biology and Chemistry, GTC Silicon Valley 2018 - ID S8827
Streaming:
Share:
 
Designing Large-Scale Machine Learning Systems with NVIDIA GPUs and Mellanox Interconnect
Come join us and learn how to build a data-centric GPU cluster for artificial intelligence. Mellanox is a leader in high-performance, scalable, low-latency network interconnects for both InfiniBand and Ethernet. We'll present the state of the art te ...Read More
Come join us and learn how to build a data-centric GPU cluster for artificial intelligence. Mellanox is a leader in high-performance, scalable, low-latency network interconnects for both InfiniBand and Ethernet. We'll present the state of the art techniques for distributed machine learning, and discuss what special requirements they impose on the system, followed by an overview of interconnect technologies used to scale and accelerate distributed machine learning including RDMA, NVIDIA's GPUDirect technology, and a special focus on the in-network computing SHARP technology used to accelerate large scale deployments in artificial intelligence and high performance computing.  Back
 
Keywords:
AI Application Deployment and Inference, Advanced AI Learning Techniques (incl. GANs and NTMs), GTC Silicon Valley 2018 - ID S8635
Streaming:
Share:
 
Accelerate Your Kaldi Speech Pipeline on the GPU
Voice commands, and advancements in automatic speech recognition algorithms, that help us interact conversationally with devices, appliances and services, are growing within our everyday environment. We will share some highlights and results from wor ...Read More
Voice commands, and advancements in automatic speech recognition algorithms, that help us interact conversationally with devices, appliances and services, are growing within our everyday environment. We will share some highlights and results from work scheduling optimizations in the Kaldi framework. The first part of the talk will describe results focused primarily on optimizing the DNN components of speech pipeline. We will then show results from a GPU optimized fast lattice decode algorithm to achieve high end to end throughput across the whole ASR pipeline from the acoustic model to the language model.  Back
 
Keywords:
AI Application Deployment and Inference, AI and DL Research, GTC Silicon Valley 2018 - ID S81034
Streaming:
Download:
Share:
 
Accelerating Large-Scale Video Surveillance for Smart Cities with TensorRT
We'll discuss a detailed scale-up method for accelerating deep learning-based object detection inference engine with INT8 by using NVIDIA's TensorRT. Previously, converting convolutional neural networks (CNNs) from 32-bit floating-point arithmetic ...Read More
We'll discuss a detailed scale-up method for accelerating deep learning-based object detection inference engine with INT8 by using NVIDIA's TensorRT. Previously, converting convolutional neural networks (CNNs) from 32-bit floating-point arithmetic (FP32) to 8-bit integer (INT8) for classification tasks has been researched. However, there is no solid work for accelerating CNN-based object detection tasks. We'll explain how to accelerate YOLO-v2, the state-of-the-art CNN-based object detector with TensorRT using INT8. We improved YOLO-v2 network for better acceleration and more accurate for surveillance and named our network SIDNet. We verified SIDNet on several benchmark object detection and intrusion detection datasets and confirmed that SIDNet with INT8 has only 1% accuracy drop compared with FP32 mode and is 5x faster than the original YOLO-v2 on NVIDIA Tesla P40.  Back
 
Keywords:
AI Application Deployment and Inference, Telecom Industry Solutions, Deep Learning and AI Frameworks, Computer Vision, Robotics & Autonomous Machines, GTC Silicon Valley 2018 - ID S8296
Streaming:
Download:
Share:
 
Applying AI to Simplify Support- Lessons Learnt
We'll provide insights into how customer support built on the foundation of AI can help streamline customer support for large enterprises, especially manufacturers. With AI technologies like image recognition and natural language processing maturing ...Read More
We'll provide insights into how customer support built on the foundation of AI can help streamline customer support for large enterprises, especially manufacturers. With AI technologies like image recognition and natural language processing maturing, enterprises should strongly consider building an AI-based support platform, especially those with an omni-channel strategy. Delivering an amazing and differentiated user experience will lead to higher net promoter and customer satisfaction scores. By employing AI-based technologies, enterprises can reduce their contacts, consequently reducing their cost and contact. It will also help them sell more replacement parts online.  Back
 
Keywords:
AI Application Deployment and Inference, NVIDIA Inception Program, Video and Image Processing, GTC Silicon Valley 2018 - ID S8517
Streaming:
Download:
Share:
 
Simulate and Validate your DNN Inference with CATIA before ADAS Industrial Deployment
One of the tough aspect of Deep Neural Network resides in its behavior validation. Although actual driving should be achieved with physical cars to train the neural network, there is today no tool to appropriately prepare data acquisition campaign or ...Read More
One of the tough aspect of Deep Neural Network resides in its behavior validation. Although actual driving should be achieved with physical cars to train the neural network, there is today no tool to appropriately prepare data acquisition campaign or go through stress validation before further on-road testing and industrial deployment. This talk will show how hardware and software in the loop on 3DEXPERIENCE CATIA, can now be extended to AI in the loop, with the ability to activate the full system engineering simulation with the actual neural network meant to run in the autonomous vehicle, accurately reproducing the neural network inference and checking overall vehicle behavior in various conditions. Every stage from full 3D synthetic data ingest and real-time software simulation, through actual hardware in the loop validation both use cases leveraging TensorRT GPU inference can now consistently be proofed for appropriate in-depth understanding of the network reactions before it drives on the road. A POC showing TensorRT and DNN behavior validation will be presented in details, opening new opportunities to validate GPU inference but also compare actual performance impact versus CPU  Back
 
Keywords:
AI Application Deployment and Inference, Product & Building Design, GTC Silicon Valley 2018 - ID S8748
Streaming:
Download:
Share:
 
Deep Learning for Industrial Inspection Analysis
We'll show how GE combines extensive domain knowledge with modern deep learning techniques to build intelligent pipeline inspection systems. GE builds a variety of industrial inspection equipment from ultrasonic pipeline inspection gauges to large-s ...Read More
We'll show how GE combines extensive domain knowledge with modern deep learning techniques to build intelligent pipeline inspection systems. GE builds a variety of industrial inspection equipment from ultrasonic pipeline inspection gauges to large-scale CT scanners. As historical producers of hardware, GE is now leading the transformation of the industrial space by building intelligent ecosystems around industrial equipment and processes. Challenges in this space include the esoteric domain-specific nature of the data, as well as the risk averse nature of the industry. However, by leveraging deep learning on large amounts of inspection data, we have been able to build a production system that enhances the reliability and consistency of the inspection process.  Back
 
Keywords:
AI Application Deployment and Inference, Industrial Inspection, GTC Silicon Valley 2018 - ID S8657
Streaming:
Share:
 
GBM Inferencing on GPU
We'll present a novel GPU implementation for batched GBM inferencing. We'll also present detailed performance comparison of our implementation against the state-of-the-art libraries such as XGBoost and Treelite. We'll then compare inference perfor ...Read More
We'll present a novel GPU implementation for batched GBM inferencing. We'll also present detailed performance comparison of our implementation against the state-of-the-art libraries such as XGBoost and Treelite. We'll then compare inference performance on various real-world datasets.  Back
 
Keywords:
AI Application Deployment and Inference, Accelerated Analytics, AI and DL Research, GTC Silicon Valley 2018 - ID S8873
Streaming:
Download:
Share:
 
How Microservices and Serverless Computing Enable the Next Generation of Machine Intelligence
We'll discuss why AI and machine learning are a natural fit for serverless computing and a general architecture for scalable and serverless machine learning in production. We'll discuss issues encountered during implementing our own on-demand scali ...Read More
We'll discuss why AI and machine learning are a natural fit for serverless computing and a general architecture for scalable and serverless machine learning in production. We'll discuss issues encountered during implementing our own on-demand scaling over GPU clusters, show how these apply to more general solutions, and present one possible vision for the future of cloud-based machine learning.  Back
 
Keywords:
AI Application Deployment and Inference, NVIDIA Inception Program, Accelerated Analytics, GTC Silicon Valley 2018 - ID S8900
Streaming:
Download:
Share:
AI Startup
Presentation
Media
Deep Learning: An Artificial Brain That Detects Any Type of Cyber Threat
Join our presentation on the first application of deep learning to cybersecurity. Deep learning is inspired by the brain's ability to learn: once a brain learns to identify an object, its identification becomes second nature. Similarly, as a ...Read More

Join our presentation on the first application of deep learning to cybersecurity. Deep learning is inspired by the brain's ability to learn: once a brain learns to identify an object, its identification becomes second nature. Similarly, as a deep learning-based artificial brain learns to detect any type of cyber threat, its prediction capabilities become instinctive. As a result, the most evasive and unknown cyber-attacks are immediately detected and prevented. We'll cover the evolution of artificial intelligence, from old rule-based systems to conventional machine learning models until current state-of-the-art deep learning models. 

  Back
 
Keywords:
AI Startup, Deep Learning and AI, GTC Silicon Valley 2017 - ID S7844
Download:
Share:
 
Disrupting Cancer Diagnostics - Cloud-based Deep Learning AI for Gigantic Pathology Images
We'll introduce a novel approach to digital pathology analytics, which brings together a powerful image server and deep learning based image analysis on a cloud platform. Recent advances in AI and Deep Learning in particular show great promi ...Read More

We'll introduce a novel approach to digital pathology analytics, which brings together a powerful image server and deep learning based image analysis on a cloud platform. Recent advances in AI and Deep Learning in particular show great promise in several fields of medicine, including pathology. Human expert judgement augmented by deep learning algorithms has the potential to speed up the diagnostic process and to make diagnostic assessments more reproducible. One of the major advantages of the novel AI-based algorithms is the ability to train classifiers for morphologies that exhibit a high level of complexity. We will present examples on context-intelligent image analysis applications, including e.g. fully automated epithelial cell proliferation assay and tumor grading. We will also present other examples of complex image analysis algorithms, which all run on-demand on whole-slide images in the cloud computing environment. Our WebMicroscope® Cloud is sold as a service (SaaS) approach, which is extremely easy to set up from a user perspective, as the need for local software and hardware installation is removed and the solution can immediately be scaled to projects of any size.

  Back
 
Keywords:
AI Startup, Healthcare and Life Sciences, Medical Imaging and Radiology, GTC Silicon Valley 2017 - ID S7856
Download:
Share:
 
Intelligent Automation using Deep Learning in Financial Services - Banking to Insurance
Long term goal of any financial institution is achieve the ability to address users with utmost experience within the boundaries of resources. It could only be a possibility when financial institutions adapt to intelligent systems. The success o ...Read More

Long term goal of any financial institution is achieve the ability to address users with utmost experience within the boundaries of resources. It could only be a possibility when financial institutions adapt to intelligent systems. The success of such systems depends heavily on the intelligence. Deep Learning has provided a huge opportunity for financial institutions to start building and planning for such large scale intelligent systems which are multi-functional and adapt. In this talk, we will discuss about how we used Deep Learning, Vega as the platform and GPUs to build high scale automation use cases in Fraud detection to complex process automation in both banking and insurance.

  Back
 
Keywords:
AI Startup, Deep Learning and AI, Finance, GTC Silicon Valley 2017 - ID S7864
Download:
Share:
AI and DL Business Track (high level)
Presentation
Media
Moving Deep Learning from Concept to Production
Spectrum Conductor with Deep Learning capabilities is an end-to-end analytics software engine for the Data Scientist, and is optimized for accelerated hardware. It's designed to support a multi-tenant, on-premises deployment for Deep Learning with a ...Read More
Spectrum Conductor with Deep Learning capabilities is an end-to-end analytics software engine for the Data Scientist, and is optimized for accelerated hardware. It's designed to support a multi-tenant, on-premises deployment for Deep Learning with and end-to-end solution means customers gain business value within each phase of the deep learning pipeline. In this session, we will explore the phases of the pipeline (Setup/Configuration, Data Preparation & Ingestion, Model Training, Deploy & Inference, and Model Maintenance) and provide insights into the unique IBM value for accelerating the use of Deep Learning across a wide variety of industries.  Back
 
Keywords:
AI and DL Business Track (high level), GTC Washington D.C. 2017 - ID DC7265
Download:
Share:
 
AI for Social Good as an Innovation Driver
Innovation can take many forms, and led by varying stakeholders across an organization. One successful model is utilizing AI for Social Good to drive a proof-of-concept that will advance a critical strategic goal. The Data Science Bowl (DSB) is ...Read More

Innovation can take many forms, and led by varying stakeholders across an organization. One successful model is utilizing AI for Social Good to drive a proof-of-concept that will advance a critical strategic goal. The Data Science Bowl (DSB) is an ideal example, launched by Booz Allen Hamilton in 2014, it galvanizes thousands of data scientists to participate in competitions that will have have far reaching impact across key industries such as healthcare. This session will explore the DSB model, as well as look at other ways organizations are utilizing AI for Social Good to create business and industry transformation.

  Back
 
Keywords:
AI and DL Business Track (high level), AI for Business, GTC Silicon Valley 2018 - ID S8953
Streaming:
Download:
Share:
 
Success in the Age of AI
From healthcare to financial services to retail, businesses are seeing unprecedented levels of efficiencies and productivity, which will only continue to rise and transform how companies operate. This session will look at how Accenture as an ent ...Read More

From healthcare to financial services to retail, businesses are seeing unprecedented levels of efficiencies and productivity, which will only continue to rise and transform how companies operate. This session will look at how Accenture as an enterprise is optimizing itself in the age of AI, as well as how it guides its customers to success. A look at best practices, insights, and measurement to help the audience inform their AI roadmap and journey.

  Back
 
Keywords:
AI and DL Business Track (high level), AI for Business, GTC Silicon Valley 2018 - ID S8984
Streaming:
Download:
Share:
 
From Dark Matter Detection to Deep Learning in Enterprise
Advancements in deep learning are enabling enterprise companies to make meaningful impacts to bottom-line profits. Enterprises capture thousands of hours of customer phone call recordings per day. This voice data is extremely valuable because it cont ...Read More
Advancements in deep learning are enabling enterprise companies to make meaningful impacts to bottom-line profits. Enterprises capture thousands of hours of customer phone call recordings per day. This voice data is extremely valuable because it contains insights that the business can use to improve customer experience and operations. We'll follow Deepgram CEO Dr. Scott Stephenson's path from working in a particle physics lab two miles underground to founding a deep learning company for voice understanding. We'll describe applications of cutting-edge AI techniques to make enterprise voice datasets mineable for valuable business insights. Companies today use these insights to drive the bottom line.  Back
 
Keywords:
AI and DL Business Track (high level), Telecom Industry Solutions, Speech and Language Processing, NVIDIA Inception Program, GTC Silicon Valley 2018 - ID S8274
Streaming:
Download:
Share:
 
The Face Will State The Case
We have all heard about Facial Expression and Recognition Systems (FERS) and emotion capture but curiosity looms large. Is it training sets born of Generative Adversarial Networks (GANs) along with GPU architectures that will catapult this technolog ...Read More
We have all heard about Facial Expression and Recognition Systems (FERS) and emotion capture but curiosity looms large. Is it training sets born of Generative Adversarial Networks (GANs) along with GPU architectures that will catapult this technology forward? To be sure, but, something much deeper - a revolution within Computer Science programs in the schools - will accelerate its arrival in consumer platforms. It's called Social Signal Processing and women technologists have a competitive advantage in inventing and enhancing the deep learning algorithms that will fuel it. Come and listen to an industry veteran with 28 years in Artificial Intelligence, including her driving Watson into consumer platforms and a graduate of Stanford University, bolstered by her solid research in Symbolic Systems, discuss their patent-pending technology in the exciting area of Social Signal Processing and FERS. They are both frequent speakers on the ethics of AI usage and will offer their thoughts about how this new class of technology offers a new deal for women to shape the future of AI.  Back
 
Keywords:
AI and DL Business Track (high level), AI and DL Research, GTC Silicon Valley 2018 - ID S8939
Streaming:
Share:
 
Matching DS Organizational Maturity to DS Skills to Optimally Grow Your Team
An organization''s data science needs change dramatically as they move through stages of data science maturity--their ability to consume, adopt, and deploy advanced analytics solutions. Understanding the maturity stage of your organization will help ...Read More
An organization''s data science needs change dramatically as they move through stages of data science maturity--their ability to consume, adopt, and deploy advanced analytics solutions. Understanding the maturity stage of your organization will help you choose projects that can bring value, grow your ability to derive greater value in the future, and help you make good decisions when growing your data science team. A data scientist might be a journeyman model builder, or a data scientist consultant, or a software engineer, or a developer of new deep learning algorithms. The data scientist that would be successful in a mature organization may well fail in an organization new to data science. Hiring and growing data scientists based on skill sets in line with your data science maturity stage and maximizes your probability of success. We''ll discuss a framework to determine your level of data science readiness, explore a tool to assess the skill sets of data scientists, and find which skills can maximize your organization''s probability of success at each stage.  Back
 
Keywords:
AI and DL Business Track (high level), GTC Silicon Valley 2018 - ID S8954
Streaming:
Download:
Share:
 
Rapid Pace of Change and Industry Progress
We are still in the early stages of AI, and its impact on industries is already significant - from healthcare to financial services to retail. Businesses are seeing unprecedented levels of efficiencies and productivity, which will only continue to ri ...Read More
We are still in the early stages of AI, and its impact on industries is already significant - from healthcare to financial services to retail. Businesses are seeing unprecedented levels of efficiencies and productivity, which will only continue to rise and transform how companies operate. This session will explore the progress of AI adoption over the last year, the industries that are leaping ahead, new AI innovations that will serve cross-industry concerns, and what businesses should expect in terms of adoption maturity in 2018.  Back
 
Keywords:
AI and DL Business Track (high level), GTC Silicon Valley 2018 - ID S8952
Streaming:
Download:
Share:
 
Scaling AI POCs Across the Enterprise
Has your team developed an AI proof-of-concept with promising metrics? Next step is to broaden the scope to impact larger areas of the enterprise. With its unique challenges and complexities, scaling POCs across multiple business units is a significa ...Read More
Has your team developed an AI proof-of-concept with promising metrics? Next step is to broaden the scope to impact larger areas of the enterprise. With its unique challenges and complexities, scaling POCs across multiple business units is a significant part of any company''s AI roadmap. This session will look at best practices, insights and success, rooted in Element AI''s experience with enterprise customers.  Back
 
Keywords:
AI and DL Business Track (high level), NVIDIA Inception Program, GTC Silicon Valley 2018 - ID S8989
Streaming:
Download:
Share:
 
Real-Time Genetic Analysis Enabled by GPU
For enterprises daunted by the prospect of AI and investing in a new technology platform, the reality is that AI can leverage already-in-place big data and cloud strategies. This session will explore AI and deep learning use cases that are desig ...Read More

For enterprises daunted by the prospect of AI and investing in a new technology platform, the reality is that AI can leverage already-in-place big data and cloud strategies. This session will explore AI and deep learning use cases that are designed for ROI, and look at how success is being measured and optimized.

  Back
 
Keywords:
AI and DL Business Track (high level), AI for Business, GTC Silicon Valley 2018 - ID S8983
Streaming:
Share:
 
The Extreme Data Economy: How Businesses Thrive in the Post Big Data Era (Presented by Kinetica)
Get the latest information on how the proliferation of mobile, cloud, and IoT devices has brought us into a new era: The Extreme Data Economy. There''s a greater variety of data than ever before, and exponentially more of it, streaming in real time. ...Read More
Get the latest information on how the proliferation of mobile, cloud, and IoT devices has brought us into a new era: The Extreme Data Economy. There''s a greater variety of data than ever before, and exponentially more of it, streaming in real time. Across industries, companies are turning data into an asset, above and beyond any product or service they offer. But unprecedented agility is required to keep business in motion and succeed in this post-big data era. To enable this level of agility, companies are turning to instant insight engines that are powered by thousands of advanced GPU cores, bringing unparalleled speed, streaming data analysis, visual foresight, and machine learning to break through the old bottlenecks. Learn about new data-powered use cases you''ll need to address, as well as advances in computing technology, particularly accelerated parallel computing, that will translate data into instant insight to power business in motion.  Back
 
Keywords:
AI and DL Business Track (high level), NVIDIA Inception Program, GTC Silicon Valley 2018 - ID S8997
Streaming:
Download:
Share:
 
Create Customer Value with Google Cloud AI (Presented by Google)
In this session, you will learn how Google Cloud helps enterprises make the most out of data, and deliver customer value. We will provide an in-depth overview of the Cloud AI and Data Analytics offering that helps enterprises manage their ML lifecycl ...Read More
In this session, you will learn how Google Cloud helps enterprises make the most out of data, and deliver customer value. We will provide an in-depth overview of the Cloud AI and Data Analytics offering that helps enterprises manage their ML lifecycle, from data ingestion to insights and prediction. We will also demonstrate some breakthrough solutions, like AutoML, that are making ML accessible to everyone.  Back
 
Keywords:
AI and DL Business Track (high level), GTC Silicon Valley 2018 - ID S8976
Streaming:
Share:
 
Trends and Opportunities for ML and AI in Consumer Insights Industries
We'll examine business value drivers for artificial intelligence and machine learning in retail and consumer goods industries. Traditionally, traction in AI and ML has been in deep research, scientific, and technical communities. Retailers and consu ...Read More
We'll examine business value drivers for artificial intelligence and machine learning in retail and consumer goods industries. Traditionally, traction in AI and ML has been in deep research, scientific, and technical communities. Retailers and consumer products companies are finding great success applying AI and ML technology to distinct use cases and business challenges. Join us to hear project descriptions and customer examples where AI and ML can impact the business by increasing revenue, protecting margin, and improving consumer satisfaction.  Back
 
Keywords:
AI and DL Business Track (high level), Virtual Reality and Augmented Reality, Consumer Engagement and Personalization, GTC Silicon Valley 2018 - ID S8131
Streaming:
Download:
Share:
 
Practical Use Cases Of AI and Deep Learning On GPUs In The Cloud For Marketing And Retail
We'll review three practical use cases of applying AI and deep learning in the marketing and retail industries. For each use case, we'll cover business situations, discuss potential approaches, and describe final solutions from both the AI and infr ...Read More
We'll review three practical use cases of applying AI and deep learning in the marketing and retail industries. For each use case, we'll cover business situations, discuss potential approaches, and describe final solutions from both the AI and infrastructural points of view. Attendees will learn about applications of AI and deep learning in marketing and advertising; AI readiness criteria; selecting the right AI and deep learning methods, infrastructure, and GPUs for specific use cases; and avoiding potential risks.  Back
 
Keywords:
AI and DL Business Track (high level), Predictive Analytics for Retail, Consumer Engagement and Personalization, GTC Silicon Valley 2018 - ID S8265
Streaming:
Download:
Share:
 
Earth Observation From Space: Deep Learning based Satellite Image Analysis
Learn how recent advances in Earth observation are opening up a new exciting area for exploration of satellite image data with deep learning. Focusing on real-world scenarios, we will teach you how to analyze this exciting remote sensing data source ...Read More
Learn how recent advances in Earth observation are opening up a new exciting area for exploration of satellite image data with deep learning. Focusing on real-world scenarios, we will teach you how to analyze this exciting remote sensing data source with deep neural networks. An automated satellite image understanding is of high interest for various research fields and industry sectors such as the insurance, agriculture or investing industry. You will learn how to apply deep neural networks in natural disaster situations and for the classification of land-use, land-cover and building types.  Back
 
Keywords:
AI and DL Business Track (high level), GIS, AI and DL Research, GTC Silicon Valley 2018 - ID S81028
Download:
Share:
AI and DL Research
Presentation
Media
Training Neural Networks with Mixed Precision: Real Examples
We will cover the techniques for training DNNs with Tensor Cores described in "S8923 - Training Neural Networks with Mixed Precision: Theory and Practice". These methods were introduced for AI processing with the Volta GPU architecture. T ...Read More
We will cover the techniques for training DNNs with Tensor Cores described in "S8923 - Training Neural Networks with Mixed Precision: Theory and Practice". These methods were introduced for AI processing with the Volta GPU architecture. Tensor Cores provide up to 120 TFlops throughput, mixing operations on IEEE half- and single-precision floats. Techniques used will include loss-scaling, master weights copy, and choosing the proper precision for a given operation. For each of TensorFlow and PyTorch we will describe a fp32 network definition and then demonstrate the same network using mixed precision techniques.  Back
 
Keywords:
AI and DL Research, Algorithms and Numerical Techniques, GTC Silicon Valley 2018 - ID S81012
Streaming:
Download:
Share:
 
Supporting a DGX Air-Gapped Production Environments
This tutorial will cover the issues encountered when deploying NVIDIA DGX-1/DGXStation into secure environment. For security reasons, some installations require that systems be isolated from the internet or outside networks. Since most DGX-1 softwar ...Read More
This tutorial will cover the issues encountered when deploying NVIDIA DGX-1/DGXStation into secure environment. For security reasons, some installations require that systems be isolated from the internet or outside networks. Since most DGX-1 software updates are accomplished through an over-the-network process with NVIDIA servers, this session will walk the participants through how updates can be made by maintaining an intermediary server. This session will be a combination of lecture, live demos and along with detailed instructions.  Back
 
Keywords:
AI and DL Research, Data Center and Cloud Infrastructure, GTC Silicon Valley 2018 - ID S8568
Streaming:
Share:
 
Scaling Machine Learning through Decentralization, Quantization, and Structured Sparsity
In this session, participants will get a taste of state-of-the-art techniques for scaling Deep Learning on GPU clusters. We present SuperML, a general and efficient communication layer for machine learning, which can scale neural network training to ...Read More
In this session, participants will get a taste of state-of-the-art techniques for scaling Deep Learning on GPU clusters. We present SuperML, a general and efficient communication layer for machine learning, which can scale neural network training to hundreds of GPU nodes. SuperML builds on three main ideas: decentralization, which allows algorithms to converge without a centralized coordinator (parameter server) or all-to-all communication, communication quantization, which significantly speeds up point-to-point messaging, and structured sparsity, by which SuperML induces model updates which only have a limited number of non-zero entries. From the technical perspective, SuperML provides a new implementation of the classic MPI standard, re-designed and re-implemented to provide efficient support for quantization and sparsity. We illustrate the performance characteristics of SuperML on CSCS Piz Daint, Europe's most powerful supercomputer, and on Amazon EC2, improving upon other highly optimized implementations such as CrayMPI and NVIDIA NCCL.  Back
 
Keywords:
AI and DL Research, Accelerated Analytics, HPC and Supercomputing, GTC Silicon Valley 2018 - ID S8668
Streaming:
Download:
Share:
 
Designing Wireless Systems with Deep Learning - An Autoencoder-Based Approach to PHY Layer Design
The field of wireless engineering is on the cusp of a revolution, driven by deep learning, that will define the next paradigm in wireless system design. While wireless communications technology has advanced considerably since its invention in the 189 ...Read More
The field of wireless engineering is on the cusp of a revolution, driven by deep learning, that will define the next paradigm in wireless system design. While wireless communications technology has advanced considerably since its invention in the 1890s, the fundamental design methodology has remained unchanged throughout its history - expert engineers hand-designing radio systems for specific applications. Deep learning enables a new, radically different approach, where systems are learned from wireless channel data. As the world becomes more connected and the Internet of Things becomes a reality, it is difficult to overstate the enormity of the impact to both commercial and military systems. This talk will provide a high-level overview of deep learning applied to wireless communications, discuss the current state of the technology and research, and present a vision for the future of wireless engineering.  Back
 
Keywords:
AI and DL Research, Telecom Industry Solutions, GTC Silicon Valley 2018 - ID S8791
Streaming:
Share:
 
Domain Adaptation Using Adversarial Training for Semantic Segmentation and Caption Style Transfer
We'll introduce the basic concept of domain adaptation and how to use adversarial training to achieve unsupervised domain adaptation. We'll then describe how the technique is used in two tasks: improving semantic segmentation across cities, and tr ...Read More
We'll introduce the basic concept of domain adaptation and how to use adversarial training to achieve unsupervised domain adaptation. We'll then describe how the technique is used in two tasks: improving semantic segmentation across cities, and transferring language style for image captioning. In particular, we combine domain adaptation with policy gradient-based reinforcement learning approach to transfer language style. The details and results of both tasks are published in ICCV 2017.  Back
 
Keywords:
AI and DL Research, GTC Silicon Valley 2018 - ID S8200
Streaming:
Share:
 
Deep Learning Applications for Radio Frequency (RF) Data
We'll discuss applications of deep learning to radio frequency (RF) data including specific signal and digital modulation scheme classification, identification of nefarious activities, and a general overview of the unique challenges and solutions fo ...Read More
We'll discuss applications of deep learning to radio frequency (RF) data including specific signal and digital modulation scheme classification, identification of nefarious activities, and a general overview of the unique challenges and solutions for AI in this domain. With the ubiquity of RF communication signals in our lives, deep learning can be leveraged to ensure accurate signal transmission and safer communities.  Back
 
Keywords:
AI and DL Research, Computational Physics, GTC Silicon Valley 2018 - ID S8826
Streaming:
Download:
Share:
 
Simultaneous Pixel-Localization and Feature Extraction for Multiple Instances in a Scene
We'll introduce attendees to a new deep learning approach to object-localization. Instead of bounding boxes, our network estimates the center pixel locations for a variable number of targets in a scene while simultaneously extracting a characteristi ...Read More
We'll introduce attendees to a new deep learning approach to object-localization. Instead of bounding boxes, our network estimates the center pixel locations for a variable number of targets in a scene while simultaneously extracting a characteristic feature-set. We'll outline the overall approach and describe the underlying network architecture and training. We'll also present the results of our network as applied to the cars overhead with context dataset and discuss the current and future possibilities of this approach.  Back
 
Keywords:
AI and DL Research, Computer Vision, GTC Silicon Valley 2018 - ID S8191
Streaming:
Share:
 
Inside NVIDIA GPU Cloud Deep Learning Framework Containers
In this technical deep dive, get an in-depth look at the deep learning containers on NVIDIA GPU Cloud (NGC) and learn how they can simplify your AI projects. NVIDIA pre-integrates and optimizes the top deep learning frameworks such as TensorFlow, PyT ...Read More
In this technical deep dive, get an in-depth look at the deep learning containers on NVIDIA GPU Cloud (NGC) and learn how they can simplify your AI projects. NVIDIA pre-integrates and optimizes the top deep learning frameworks such as TensorFlow, PyTorch, and MXNet, and makes them available on NVIDIA GPU Cloud, removing time consuming do-it-yourself software integration. We'll look at the NVIDIA framework optimizations, such as reducing GPU memory overhead, improving multi-GPU scaling, and reducing latency. And we'll talk about the integration of runtimes and drivers in the containers to ensure the correct versions of software are working together for peak performance. You'll leave with an understanding of what make an NVIDIA GPU-optimized deep learning container tick.  Back
 
Keywords:
AI and DL Research, Deep Learning and AI Frameworks, Data Center and Cloud Infrastructure, GTC Silicon Valley 2018 - ID S8497
Streaming:
Download:
Share:
 
Matchbox: Automatic Batching for Dynamic Deep Learning
Matchbox is an open source PyTorch-based tool that lets users implement their deep learning models as imperative code that applies to individual data samples, then efficiently train and validate them on batched data using GPUs. By automatically keepi ...Read More
Matchbox is an open source PyTorch-based tool that lets users implement their deep learning models as imperative code that applies to individual data samples, then efficiently train and validate them on batched data using GPUs. By automatically keeping track of batch-level masking and padding and rewriting data-dependent control flow, Matchbox simplifies model code, eliminates a class of implementation bugs, and allows programmers to work directly at a more natural level of abstraction.  Back
 
Keywords:
AI and DL Research, Deep Learning and AI Frameworks, GTC Silicon Valley 2018 - ID S8977
Streaming:
Download:
Share:
 
Tackling the Crowded Radio Frequency Spectrum Using Deep Learning
We'll introduce new concepts and algorithms that apply deep learning to radio frequency (RF) data to advance the state of the art in signal processing and digital communications. With the ubiquity of wireless devices, the crowded RF spectrum ...Read More

We'll introduce new concepts and algorithms that apply deep learning to radio frequency (RF) data to advance the state of the art in signal processing and digital communications. With the ubiquity of wireless devices, the crowded RF spectrum poses challenges for cognitive radio and spectral monitoring applications. Furthermore, the RF modality presents unique processing challenges due to the complex-valued data representation, large data rates, and unique temporal structure. We'll present innovative deep learning architectures to address these challenges, which are informed by the latest academic research and our extensive experience building RF processing solutions. We'll also outline various strategies for pre-processing RF data to create feature-rich representations that can significantly improve performance of deep learning approaches in this domain. We'll discuss various use-cases for RF processing engines powered by deep learning that have direct applications to telecommunications, spectral monitoring, and the Internet of Things.

  Back
 
Keywords:
AI and DL Research, Telecom Industry Solutions, Federal, GTC Silicon Valley 2018 - ID S8267
Streaming:
Download:
Share:
 
Point Cloud Deep Learning
This presentation shows in-depth comparisons of several neural network models for 3D object classification. Object classification from 2D image is studied thoroughly and widely adopted during last few years by following the advances of deep neural ne ...Read More
This presentation shows in-depth comparisons of several neural network models for 3D object classification. Object classification from 2D image is studied thoroughly and widely adopted during last few years by following the advances of deep neural networks. From then, 3D object classification methods are actively studied, and yet not completely mature. Point cloud is most basic format of 3D objects. In this work, we present many neural network models that can be learned from 3D point cloud. It includes directly learning from 3D point cloud, projected 2D pixels, and voxelated volumes. This work uses Princeton ModelNet datasets and ShapeNetCore.v2 dataset, and then provides the comparisons of those neural network models.  Back
 
Keywords:
AI and DL Research, Graphics and AI, Rendering and Ray Tracing, Real-Time Graphics, GTC Silicon Valley 2018 - ID S8453
Streaming:
Download:
Share:
 
GUNREAL: GPU-Accelerated Unsupervised Reinforcement and Auxiliary Learning
We'll introduce GPU-accelerated unsupervised reinforcement and auxiliary learning (UNREAL) algorithm. Recent state-of-the-art deep reinforcement learning algorithms, such as A3C and UNREAL, are designed to train on a single device with only CPUs. Us ...Read More
We'll introduce GPU-accelerated unsupervised reinforcement and auxiliary learning (UNREAL) algorithm. Recent state-of-the-art deep reinforcement learning algorithms, such as A3C and UNREAL, are designed to train on a single device with only CPUs. Using GPU acceleration for these algorithms results in low GPU utilization, which means the full performance of the GPU is not reached. Motivated by the architecture changes made by the GA3C algorithm, which gave A3C better GPU acceleration, together with the high learning efficiency of the UNREAL algorithm, we extend GA3C with the auxiliary tasks from UNREAL to create GUNREAL. We show that our GUNREAL system finished training faster than UNREAL and reached higher scores than GA3C.  Back
 
Keywords:
AI and DL Research, Performance Optimization, GTC Silicon Valley 2018 - ID S8219
Streaming:
Download:
Share:
 
Large-Scale Self-Supervised Robot Learning with GPU-Enabled Video-Prediction Models
To acquire rich repertoires of skills, robots must be able to learn from their own autonomously collected data. We'll describe a video-prediction model that predicts what a robot will see next, and show how this model can be used to solve complex ma ...Read More
To acquire rich repertoires of skills, robots must be able to learn from their own autonomously collected data. We'll describe a video-prediction model that predicts what a robot will see next, and show how this model can be used to solve complex manipulations tasks in real-world settings. Our model was trained on 44,000 video sequences, where the manipulator autonomously pushes various objects. Using the model, the robot is capable of moving objects that were not seen during training to desired locations, handling multiple objects and pushing objects around obstructions. Unlike other methods in robotic learning, video-prediction does not require any human labels. Our experiments show that the method achieves a significant advance in the range and complexity of skills that can be performed entirely with self-supervised robotic learning. This session is for attendees that possess a basic understanding of convolutional and recurrent neural networks.  Back
 
Keywords:
AI and DL Research, IoT, Robotics & Drones, Robotics & Autonomous Machines, GTC Silicon Valley 2018 - ID S8629
Streaming:
Download:
Share:
 
Deep Generative Modeling for Speech Synthesis and Sensor Data Augmentation
We'll discuss how we could use deep generative modeling in two application domains; in speech synthesis, and in sensor data modeling. We'll give an overview of what generative modeling is and how it could be used for practical AI tasks through the ...Read More
We'll discuss how we could use deep generative modeling in two application domains; in speech synthesis, and in sensor data modeling. We'll give an overview of what generative modeling is and how it could be used for practical AI tasks through these examples. We'll also give a flavor of latent space methods, which we can use to learn more about our data so as to transform them in meaningful ways, with uses in both reconstruction and in generation.  Back
 
Keywords:
AI and DL Research, Advanced AI Learning Techniques (incl. GANs and NTMs), GTC Silicon Valley 2018 - ID S8617
Streaming:
Download:
Share:
 
New Applications of Deep Learning in Dialogue Generation and Question Answering
The current generation AI systems are mostly moving towards dialogue generation and question answering. Human like conversation and dialogue based interaction has been proposed as the interface for tomorrow, which would obliterate key-boards and trac ...Read More
The current generation AI systems are mostly moving towards dialogue generation and question answering. Human like conversation and dialogue based interaction has been proposed as the interface for tomorrow, which would obliterate key-boards and track-pads from computers as we know them. We present two important current developments in these fields. First we talk about a neural dialogue generation system which can be deployed to engage humans in a multi-turn conversation. Next we talk about a segmented question answering module which can find answers from the web. The combination of these two techniques has the potential to unlock numerous new verticals, such as travel, retail etc. We will talk about the technical details as well as the higher level design choices.  Back
 
Keywords:
AI and DL Research, Speech and Language Processing, Advanced AI Learning Techniques (incl. GANs and NTMs), GTC Silicon Valley 2018 - ID S8151
Streaming:
Share:
 
Object-Level Deep Reinforcement Learning
We'll show how deep reinforcement learning can be greatly sped up by separating perception and action, with a reward function specified in terms of objects and their motions, which are supplied by the perceptual system. In the past five years, reinf ...Read More
We'll show how deep reinforcement learning can be greatly sped up by separating perception and action, with a reward function specified in terms of objects and their motions, which are supplied by the perceptual system. In the past five years, reinforcement learners have become vastly more powerful by incorporating deep learning techniques, playing Atari, Mario, Go, and other games with superhuman skill. However, these learners require vast amounts of training data to become skilled. For example, to master Pong, state-of-the-art reinforcement learners require tens of millions of game frames, equivalent to months of play time at human speed. We show that endowing the learner with a minimal perceptual system, capable of detecting and tracking objects, greatly reduces the number of frames needed for learning. This shifts the learning bottleneck from the amount of training data available to computations easily accelerated with GPUs.  Back
 
Keywords:
AI and DL Research, Advanced AI Learning Techniques (incl. GANs and NTMs), GTC Silicon Valley 2018 - ID S8581
Streaming:
Download:
Share:
 
Recent Advances in Neural Machine Translation: Multilingual, Non-Parametric to Unsupervised Neural Machine Translation
We'll describe the latest advances in neural machine translation from three different perspectives. We'll start with character-level, multilingual neural machine translation, which aims at harnessing positive language transfer among multiple langua ...Read More
We'll describe the latest advances in neural machine translation from three different perspectives. We'll start with character-level, multilingual neural machine translation, which aims at harnessing positive language transfer among multiple languages to improve the translation quality and the robustness of such a multilingual translation model to intra-sentence code-switching and typos. We'll then discuss the recent research on exploiting data beside oft-used parallel corpora. We'll discuss how another modality, such as vision, can be used to enable zero-resource machine translation, and how purely unsupervised neural machine translation can be done by exploiting the similarity between language distributions of two languages. Finally, we'll discuss a recent trend of retrieval-based approaches to deep learning with a specific example of non-parametric neural machine translation.  Back
 
Keywords:
AI and DL Research, Advanced AI Learning Techniques (incl. GANs and NTMs), GTC Silicon Valley 2018 - ID S8609
Streaming:
Download:
Share:
 
Deep Active Learning
We'll discuss ongoing work at NVIDIA on deep active learning. Attendees can expect to learn what active learning is and some of the challenges of applying it to deep neural network training.
We'll discuss ongoing work at NVIDIA on deep active learning. Attendees can expect to learn what active learning is and some of the challenges of applying it to deep neural network training.  Back
 
Keywords:
AI and DL Research, Advanced AI Learning Techniques (incl. GANs and NTMs), GTC Silicon Valley 2018 - ID S8692
Streaming:
Share:
 
Unsupervised Image-to-Image Translation Networks
We'll introduce a GAN-based framework for unsupervised image-to-image translation. It leverages a shared latent space assumption to learn to translate an image in one domain to a corresponding image in another domain without requiring any pair of co ...Read More
We'll introduce a GAN-based framework for unsupervised image-to-image translation. It leverages a shared latent space assumption to learn to translate an image in one domain to a corresponding image in another domain without requiring any pair of corresponding images in the two domains in the training dataset. We'll show examples on translating street scene images, from sunny day to rainy day or from day time to night time. We also show image translation results on dog breed conversions and cat species conversion as well as human face translation based on attributes.  Back
 
Keywords:
AI and DL Research, Computer Vision, GTC Silicon Valley 2018 - ID S8114
Streaming:
Download:
Share:
 
Towards Lifelong Reinforcement Learning
Reinforcement learning aims to determine a mapping from observations to actions that maximize a reward criterion. The agent starts off exploring the environment for rewards with random search, which is only likely to succeed in all but simplest of se ...Read More
Reinforcement learning aims to determine a mapping from observations to actions that maximize a reward criterion. The agent starts off exploring the environment for rewards with random search, which is only likely to succeed in all but simplest of settings. Furthermore, measuring and designing reward functions for real-world tasks is non-trivial. Inspired by research in developmental psychology, in this talk I will discuss how reinforcement learning agents might use curiosity and knowledge accumulated from experience for efficient exploration. I will present results illustrating an agent learning to play the game of Mario and learning to navigate without rewards, a study quantifying the kinds of prior knowledge used by humans for efficient exploration and some robotic manipulation experiments including the use of an anthropomorphic hand for grasping objects.   Back
 
Keywords:
AI and DL Research, IoT, Robotics & Drones, Robotics & Autonomous Machines, GTC Silicon Valley 2018 - ID S8217
Streaming:
Share:
 
How We Can Analyze Profile from Real-Time Conversation by Unsupervised Learning
To convert phonemes of telephone conversations and responses at meetings into texts in real time, pass the text to the computational model created by DGX-1, label with a learning without teacher, and add the clusters, we are developing a system which ...Read More
To convert phonemes of telephone conversations and responses at meetings into texts in real time, pass the text to the computational model created by DGX-1, label with a learning without teacher, and add the clusters, we are developing a system which compares objects and analyzes meaning of conversation and profiles of interlocutors. With this technology, customers can receive appropriate responses at the beginning of a conversation with a help desk, and patients can receive correspondence during a remote diagnosis with a doctor based solely off of their dialogue and examination results. By using TensorFlow as a platform and running the K-Means method, Word2vec, Doc2Vec, etc. in DGX-1 clustered environment on DGX-1, the result of arithmetic processing is found at high speed conversation. Even if the amount of sentences is increased, the learning effect increases linearly, demonstrating that the proportion of validity can be raised without taking grammar of languages ??other than English (e.g. Japanese) into account.  Back
 
Keywords:
AI and DL Research, Speech and Language Processing, NVIDIA Inception Program, GTC Silicon Valley 2018 - ID S8371
Streaming:
Download:
Share:
 
Embodied Question Answering
Building intelligent agents that possess the ability to perceive the rich visual environment around us, communicate this understanding in natural language to humans and other agents, and execute actions in a physical environment, has been a long-term ...Read More
Building intelligent agents that possess the ability to perceive the rich visual environment around us, communicate this understanding in natural language to humans and other agents, and execute actions in a physical environment, has been a long-term goal of Artificial Intelligence. In this talk, I will present my recent work on an instantiation of this goal -- Embodied Question Answering (EQA) -- where an agent that is spawned at a random location in an environment (a house or building) is asked a natural language question ("What color is the car?"). The agent perceives its environment through first-person vision and can perform a few 'atomic' actions: move-{forward, backward, right, left}, and turn-{right, left}. The objective of the agent is to explore the environment and gather visual information necessary to answer the question ("orange"). I'll introduce our OpenGL-based environments, a large-scale dataset of expert demonstrations for this task and deep models, trained end-to-end using reinforcement learning, from raw pixels to multi-step navigation control to visual question answering.  Back
 
Keywords:
AI and DL Research, Computer Vision, GTC Silicon Valley 2018 - ID S8582
Streaming:
Download:
Share:
 
Meet Horovod: Uber's Open Source Distributed Deep Learning Framework for TensorFlow
Horovod makes it easy to train a single GPU TensorFlow model on many GPUs; both on a single server and across multiple servers. We'll cover Uber's explorations of distributed deep learning, how to use Horovod, and what kind of performance you ...Read More
Horovod makes it easy to train a single GPU TensorFlow model on many GPUs; both on a single server and across multiple servers. We'll cover Uber's explorations of distributed deep learning, how to use Horovod, and what kind of performance you can get on standard models, such as Inception V3 and ResNet-101. Learn how to speed up training of your TensorFlow model with Horovod.  Back
 
Keywords:
AI and DL Research, Deep Learning and AI Frameworks, HPC and AI, GTC Silicon Valley 2018 - ID S8152
Streaming:
Share:
 
Instance-Aware Image and Sentence Matching with Selective Multimodal LSTM
We'll present a unique framework for cross-modal image and sentence matching; namely selective multimodal long short-term memory (LSTM) that incorporates a new deep learning module as multimodal context-modulated attention network to selectively att ...Read More
We'll present a unique framework for cross-modal image and sentence matching; namely selective multimodal long short-term memory (LSTM) that incorporates a new deep learning module as multimodal context-modulated attention network to selectively attend to pairwise semantic concepts. In detail, effective image and sentence matching depends on measuring their global visual-semantic similarity. Based on the observation that such a global similarity arises from a complex aggregation of multiple local similarities between pairwise instances of image (objects) and sentence (words), we propose a selective multimodal LSTM network (sm-LSTM) for instance-aware image and sentence matching. The sm-LSTM includes a multimodal context-modulated attention scheme at each timestep that can selectively attend to a pair of instances of image and sentence by predicting pairwise instance-aware saliency maps for image and sentence. For selected pairwise instances, their representations are obtained based on the predicted saliency maps, and then compared to measure their local similarity. By similarly measuring multiple local similarities within a few timesteps, the sm-LSTM sequentially aggregate.  Back
 
Keywords:
AI and DL Research, Computer Vision, GTC Silicon Valley 2018 - ID S8281
Streaming:
Download:
Share:
 
Towards AI Agents That Can See, Talk, and Act
We are witnessing unprecedented advances in computer vision and AI. What lies next for AI? We believe that the next generation of intelligent systems (say the next generation of Google's Assistant, Facebook's M, Apple's Siri, Amazon's Alexa) will ...Read More
We are witnessing unprecedented advances in computer vision and AI. What lies next for AI? We believe that the next generation of intelligent systems (say the next generation of Google's Assistant, Facebook's M, Apple's Siri, Amazon's Alexa) will need to possess the ability to perceive their environment (through vision, audition, or other sensors), communicate (i.e., hold a natural language dialog with humans and other agents), and act (e.g., aid humans by executing API calls or commands in a virtual or embodied environment), for tasks such as: aiding visually impaired users in understanding their surroundings; interacting with an AI assistant (Human: 'Alexa can you see the baby in the baby monitor?', AI: 'Yes, I can', Human: 'Is he sleeping or playing?'); robotics applications (e.g. search and rescue missions) where the operator may be situationally blind and operating via language. We'll present work from our lab on a range of projects on such visually grounded conversational agents.  Back
 
Keywords:
AI and DL Research, Computer Vision, GTC Silicon Valley 2018 - ID S8571
Streaming:
Download:
Share:
 
Scaling Convolutional Neural Networks with Kubernetes and TensorFlow on AWS GPUs
In this session we present a Kubernetes deployment on Amazon AWS GPUs that provide customized computer vision to a large number of users. Reza offers an overview of Matroid's pipeline and demonstrates how to customize computer vision neural network ...Read More
In this session we present a Kubernetes deployment on Amazon AWS GPUs that provide customized computer vision to a large number of users. Reza offers an overview of Matroid's pipeline and demonstrates how to customize computer vision neural network models in the browser, followed by building, training, and visualizing TensorFlow models, which are provided at scale to monitor video streams.  Back
 
Keywords:
AI and DL Research, Data Center and Cloud Infrastructure, Computer Vision, GTC Silicon Valley 2018 - ID S8610
Streaming:
Download:
Share:
 
Audio Recognition, Context-Awareness, and its Applications
We'll explain the concept and the importance of audio recognition, which aims to understand literally all the information contained in the audio, not limiting its scope to speech recognition. It includes the introduction of various types of non ...Read More
We'll explain the concept and the importance of audio recognition, which aims to understand literally all the information contained in the audio, not limiting its scope to speech recognition. It includes the introduction of various types of non-verbal information contained in the audio such as acoustic scenes/events, speech, and music. This session is helpful to the people who are not familiar with audio processing but are interested in the context-aware system. Also, it might be inspiring for someone who develops AI applications such as AI home assistant, a humanoid robot, and self-driving cars. It also covers the potential use-cases and creative applications, including a video demonstration of the audio context-aware system applied to media-art performance for real-time music generation.  Back
 
Keywords:
AI and DL Research, Speech and Language Processing, NVIDIA Inception Program, GIS, GTC Silicon Valley 2018 - ID S8696
Streaming:
Download:
Share:
 
Trade and Manage Wealth with Deep Reinforcement Learning and Memory
We'll present how deep reinforcement learning (DRL) and memory extended networks can be used to train agents, which optimize asset allocations or propose trading actions. The memory component is crucial for improved mini-batch parallelization and he ...Read More
We'll present how deep reinforcement learning (DRL) and memory extended networks can be used to train agents, which optimize asset allocations or propose trading actions. The memory component is crucial for improved mini-batch parallelization and helps mitigate catastrophic forgetting. We also address how concepts from risk-sensitive and safe reinforcement learning apply to improve the robustness of the learned policies. The DRL approach has several advantages over the industry standard approach, which is still based on the mean variance portfolio optimization. The most significant benefit is that the information bottleneck between the statistical return model and the portfolio optimizer is removed, and available market data and trade history are used much more efficiently.  Back
 
Keywords:
AI and DL Research, Algorithms and Numerical Techniques, Advanced AI Learning Techniques (incl. GANs and NTMs), Finance, GTC Silicon Valley 2018 - ID S8679
Streaming:
Download:
Share:
 
(Deep) Learning to Grasp with a Close-Loop DNN Controller
The paradigm for robot programming is changing with the adoption of the deep learning approach in the field of robotics. Instead of hard coding a complex sequence of actions, tasks are acquired by the robot through an active learning procedure. This ...Read More
The paradigm for robot programming is changing with the adoption of the deep learning approach in the field of robotics. Instead of hard coding a complex sequence of actions, tasks are acquired by the robot through an active learning procedure. This introduces new challenges that have to be solved to achieve effective training. We'll show several issues that can be encountered while learning a close-loop DNN controller aimed at a fundamental task like grasping, and their practical solutions. First, we'll illustrate the advantages of training using a simulator, as well as the effects of choosing different learning algorithms in the reinforcement learning and imitation learning domains. We'll then show how separating the control and vision modules in the DNN can simplify and speed up the learning procedure in the simulator, although the learned controller hardly generalizes to the real world environment. Finally, we'll demonstrate how to use domain transfer to train a DNN controller in a simulator that can be effectively employed to control a robot in the real world.  Back
 
Keywords:
AI and DL Research, IoT, Robotics & Drones, Computer Vision, Robotics & Autonomous Machines, GTC Silicon Valley 2018 - ID S8132
Streaming:
Download:
Share:
 
Affective Categorization Using Contactless-Based Accelerometers
We'll cover the four known methods for emotion detection: vision, speech, sentiment analysis, and wearable technology. We'll provide a quick dive through each presented solution, and then introduce a novel approach aimed for the future of autonomou ...Read More
We'll cover the four known methods for emotion detection: vision, speech, sentiment analysis, and wearable technology. We'll provide a quick dive through each presented solution, and then introduce a novel approach aimed for the future of autonomous vehicles.  Back
 
Keywords:
AI and DL Research, Consumer Engagement and Personalization, GTC Silicon Valley 2018 - ID S8352
Streaming:
Download:
Share:
 
Graduate Fellowship FastForward Talks
Join a special presentation from our 2017-2018 Graduate Fellowship recipients to learn "what's next" out of the world of research and academia. Sponsored projects involve a variety of technical challenges, including distributed systems for ...Read More
Join a special presentation from our 2017-2018 Graduate Fellowship recipients to learn "what's next" out of the world of research and academia. Sponsored projects involve a variety of technical challenges, including distributed systems for large-scale deep learning; dynamic data structures for massively parallel machine learning; machine learning techniques for biomedical image analysis; visual dynamics; and compilation frameworks for high-performance graphics systems. We believe that these minds lead the future in our industry and we're proud to support the 2016-2017 NVIDIA Graduate Fellows. We'll also announce the 2017-2018 Graduate Fellows at this session. For more information on the NVIDIA Graduate Fellowship program, visit www.nvidia.com/fellowship.  Back
 
Keywords:
AI and DL Research, Virtual Reality and Augmented Reality, Graphics and AI, Computational Biology and Chemistry, Computer Vision, GTC Silicon Valley 2018 - ID S8793
Streaming:
Share:
 
Learning Rigidity in Dynamic Scenes for Scene Flow Estimation
Estimation of 3D motion in a dynamic scene from a pair of images is a core task in many scene understanding problems. In real world applications, a dynamic scene is commonly captured by a moving camera (i.e., panning, tilting or hand-held), increasin ...Read More
Estimation of 3D motion in a dynamic scene from a pair of images is a core task in many scene understanding problems. In real world applications, a dynamic scene is commonly captured by a moving camera (i.e., panning, tilting or hand-held), increasing the task complexity because the scene is observed from different viewpoints. The main challenge is the disambiguation of the camera motion from scene motions, which becomes more difficult as the amount of rigid parts observed decreases. In this talk, We introduce a method to learn a rigidity of a scene from a large collection of dynamic scene data, and directly infer a rigidity mask from two sequential RGB-D images in a supervised manner. With the learned network, we show how we can effectively estimate camera motion and projected scene flow using computed 2D optical flow and the inferred rigidity mask. Through evaluations, we show that our methods can make the scene flow estimation more robust and stable over state-of-the-art methods in challenging dynamic scenes. The expected audiences will include people who are interested in computer vision algorithms, but not limited to any audiences interested in AI and machine learning in general. We'll cover: the motivation behind scene flow estimation, potential applications, how we train two networks for the scene flow estimation, and how we evaluate the algorithm with popular benchmark dataset, SINTEL. We'll also show a new semi-synthetic dataset and its generation method where we mix real video footage with virtually rendered foreground scenes.  Back
 
Keywords:
AI and DL Research, Computer Vision, GTC Silicon Valley 2018 - ID S8798
Streaming:
Download:
Share:
 
Deep Learning for Transportation: Fast Estimation of Travel Times Using Historical Routes
During this presentation we will review a deep neural network architecture and its training approaches used for producing high volume of estimations of travel times on a road graph with historical routes and traffic. This includes initial and continu ...Read More
During this presentation we will review a deep neural network architecture and its training approaches used for producing high volume of estimations of travel times on a road graph with historical routes and traffic. This includes initial and continuous online training, finding various sources to produce training data, challenges of quality control, and, of course, the invaluable role of GPU's for computation during both training and inference.  Back
 
Keywords:
AI and DL Research, Product & Building Design, Intelligent Video Analytics and Smart Cities, GIS, Autonomous Vehicles, GTC Silicon Valley 2018 - ID S8156
Streaming:
Download:
Share:
 
Block-Sparse Recurrent Neural Networks
Recurrent neural networks are used in state-of-the-art models in domains such as speech recognition, machine translation, and language modeling. Sparsity is a technique to reduce compute and memory requirements of deep learning models. Sparse RNNs ar ...Read More
Recurrent neural networks are used in state-of-the-art models in domains such as speech recognition, machine translation, and language modeling. Sparsity is a technique to reduce compute and memory requirements of deep learning models. Sparse RNNs are easier to deploy on devices and high-end server processors. Even though sparse operations need less compute and memory relative to their dense counterparts, the speed-up observed by using sparse operations is less than expected on different hardware platforms. To address this issue, we prune blocks of weights in a layer instead of individual weights. Using these techniques, we can create block-sparse RNNs with sparsity ranging from 80% to 90% with a small loss in accuracy. This technique allows us to reduce the model size by 10x. Additionally, we can prune a larger dense network to recover this loss in accuracy while maintaining high block sparsity and reducing the overall parameter count. Our technique works with a variety of block sizes up to 32x32. Block-sparse RNNs eliminate overheads related to data storage and irregular memory accesses while increasing hardware efficiency compared to unstructured sparsity.  Back
 
Keywords:
AI and DL Research, HPC and AI, GTC Silicon Valley 2018 - ID S8924
Streaming:
Download:
Share:
 
Learning from Limited Data
Constructing an accurate prediction model from limited data is one of the important tasks in machine learning. We'll introduce unsupervised domain adaptation and a learning method using interclass patterns as a method to construct accurate predictio ...Read More
Constructing an accurate prediction model from limited data is one of the important tasks in machine learning. We'll introduce unsupervised domain adaptation and a learning method using interclass patterns as a method to construct accurate prediction models from limited data. Regarding unsupervised domain adaptation, we use three networks asymmetrically. Two networks are used to label unlabeled target patterns, and one network is trained by the pseudo-labeled patterns to obtain target-discriminative representations. About the learning method using interclass patterns, we generate interclass patterns by mixing two patterns belonging to different classes with a random ratio and train the model to output the mixing ratio form the mixed patterns. Although the algorithm is very simple, the proposed method significantly improves classification performance on sound recognition and image recognition. In addition, we'll briefly introduce various topics, including WebDNN, which our team is working on.  Back
 
Keywords:
AI and DL Research, GTC Silicon Valley 2018 - ID S8786
Streaming:
Download:
Share:
 
Deep Generative Models for Image and Video Creation
We'll focus on recent developments in deep learning-based generative models for image and video creation. The last two to three years have seen an explosive growth in the development of generative adversarial networks, variational autoencoders, and ...Read More
We'll focus on recent developments in deep learning-based generative models for image and video creation. The last two to three years have seen an explosive growth in the development of generative adversarial networks, variational autoencoders, and related autoregressive methods that have been made it possible to automatically generate images and videos, by harnessing the power of GPUs and deep learning libraries. These methods present interesting possibilities in automatic generation of datasets for training machine learning methods, as well as in real-world applications for image and video processing such as morphing, editing, advertising, design, and art. We'll present the technical details of these methods and recent results in various settings.  Back
 
Keywords:
AI and DL Research, Advanced AI Learning Techniques (incl. GANs and NTMs), Video and Image Processing, GTC Silicon Valley 2018 - ID S8784
Streaming:
Share:
 
Geometry-Aware Learning of Maps for Camera Localization
Maps are a key component in image-based camera localization and visual SLAM systems: they are used to establish geometric constraints between images, correct drift in relative pose estimation, and relocalize cameras after lost tracking. The exact def ...Read More
Maps are a key component in image-based camera localization and visual SLAM systems: they are used to establish geometric constraints between images, correct drift in relative pose estimation, and relocalize cameras after lost tracking. The exact definitions of maps, however, are often application-specific and hand-crafted for different scenarios (e.g., 3D landmarks, lines, planes, bags of visual words). We propose to represent maps as a deep neural net called MapNet, which enables learning a data-driven map representation. Unlike prior work on learning maps, MapNet exploits cheap and ubiquitous sensory inputs like visual odometry and GPS in addition to images and fuses them together for camera localization. Geometric constraints expressed by these inputs, which have traditionally been used in bundle adjustment or pose-graph optimization, are formulated as loss terms in MapNet training and also used during inference. In addition to directly improving localization accuracy, this allows us to update the MapNet (i.e., maps) in a self-supervised manner using additional unlabeled video sequences from the scene.  Back
 
Keywords:
AI and DL Research, Autonomous Vehicles, Computer Vision, GTC Silicon Valley 2018 - ID S8792
Streaming:
Share:
 
Dense Connection Networks for Conversational Speech Recognition
Densely connected neural networks were originally introduced to avoid the problem of layer-wise vanishing gradients when CNNs are stacked in a very deep fashion, specifically for image recognition tasks. Inspired by these works, we've explored the u ...Read More
Densely connected neural networks were originally introduced to avoid the problem of layer-wise vanishing gradients when CNNs are stacked in a very deep fashion, specifically for image recognition tasks. Inspired by these works, we've explored the use of dense networks connections within LSTM models for the task of automatic speech recognition. By introducing additional connections, to connect (almost) every layer to at least one other layer, we mitigate the vanishing gradient effect between LSTM layers and enable error signals to propagated back to the very first layer during training. In this presentation, we'll present the fundamentals of speech recognition and introduce different neural network model structures that have been shown to be effective for this task. We'll then introduce identity, highway, and dense connections and demonstrate how they improve the performance of these models. We'll evaluate the performance of these models across different datasets, and show that with a lattice-based system combination, densely connected LSTMs significantly contributed to reaching the marks of 5.0% and 9.1% in word error rate (WER) for the Switchboard and CallHome testsets.  Back
 
Keywords:
AI and DL Research, Speech and Language Processing, GTC Silicon Valley 2018 - ID S8903
Streaming:
Share:
 
Deep Learning Applications in E-Commerce
In this talk we will present four applications of deep learning in e-commerce. 1) A deep neural net architecture which has been successfully deployed as a large scale Visual Search and Recommendation system for e-commerce. The deployment has been at ...Read More
In this talk we will present four applications of deep learning in e-commerce. 1) A deep neural net architecture which has been successfully deployed as a large scale Visual Search and Recommendation system for e-commerce. The deployment has been at Flipkart, India's largest e-Commerce vendor, over a catalog of 50M products, supporting 2K queries per second. Our results beat state of the art on the on the Exact Street2Shop dataset. 2) Visual Semantic embedding of e-Commerce products for enhanced searchability and product ranking. 3) Neural Network based click prediction. 4) A novel neural network architecture for demand prediction.  Back
 
Keywords:
AI and DL Research, Deep Learning and AI Frameworks, Consumer Engagement and Personalization, Computer Vision, GTC Silicon Valley 2018 - ID S8684
Streaming:
Share:
 
Model Architectures and Training Techniques for High-Precision Landmark Localization
We'll discuss training techniques and deep learning architectures for high-precision landmark localization. In the first part of the session, we'll talk about ReCombinator Networks, which aims at maintaining pixel-level image information ...Read More

We'll discuss training techniques and deep learning architectures for high-precision landmark localization. In the first part of the session, we'll talk about ReCombinator Networks, which aims at maintaining pixel-level image information, for high-accuracy landmark localization. This model combines coarse-to-fine features to first observe global (coarse) image information and then recombines local (fine) information. By using this model, we report SOTA on three facial landmark datasets. This model can be used for other tasks that require pixel-level accuracy (for example, image segmentation, image-to-image translation). In the second part, we'll talk about improving landmark localization in a semi-supervised setting, where less labeled data is provided. Specifically, we consider a scenario where few labeled landmarks are given during training, but lots of weaker labels (for example, face emotions, hand gesture) that are easier to obtain are provided. We'll describe training techniques and model architectures that can leverage weaker labels to improve landmark localization.

  Back
 
Keywords:
AI and DL Research, Computer Vision, GTC Silicon Valley 2018 - ID S8406
Streaming:
Download:
Share:
 
Learning Robotic Plans from Real-World Demonstrations Using only Randomized Simulated Images
Using only randomized simulated images, we'll present a system to infer and simply execute a human-readable robotic program after watching a real-world task demonstration. The system is comprised of a series of deep neural network modules, each lear ...Read More
Using only randomized simulated images, we'll present a system to infer and simply execute a human-readable robotic program after watching a real-world task demonstration. The system is comprised of a series of deep neural network modules, each learned entirely in simulation. During training, images are generated in a gaming engine and made transferable to the real world by domain randomization. After training, the system is straightforwardly deployed on a real robot with no retuning of the neural networks and having never previously seen a real image. We demonstrate the system on a Baxter robot performing block tower construction tasks.  Back
 
Keywords:
AI and DL Research, IoT, Robotics & Drones, Robotics & Autonomous Machines, GTC Silicon Valley 2018 - ID S8439
Streaming:
Share:
 
Can AI Generate Love Advice? Neural Conclusion-Supplement Answer Generation for Non Factoid Questions
Learn how to generate long answers for non-factoid questions in quality assurance community sites by using the encoder-decoder framework. We'll present our novel extension of the encoder-decoder framework, called the ensemble network, that goes beyo ...Read More
Learn how to generate long answers for non-factoid questions in quality assurance community sites by using the encoder-decoder framework. We'll present our novel extension of the encoder-decoder framework, called the ensemble network, that goes beyond a single short sentence. It handles several sentences (i.e. two major sentence types that organize answers for non-factoid questions, conclusion statements, and its supplementary ones) to generate complicated non-factoid answers.  Back
 
Keywords:
AI and DL Research, Speech and Language Processing, GTC Silicon Valley 2018 - ID S8301
Streaming:
Share:
 
Re3: Realtime Recurrent Regression Networks for Visual Tracking of Generic Objects
Robust object tracking requires knowledge and understanding of the object being tracked: its appearance, motion, and change over time. A tracker must be able to modify its underlying model and adapt to new observations. We present Re3, a real-ti ...Read More

Robust object tracking requires knowledge and understanding of the object being tracked: its appearance, motion, and change over time. A tracker must be able to modify its underlying model and adapt to new observations. We present Re3, a real-time deep object tracker capable of incorporating temporal information into its model. Rather than focusing on a limited set of objects or training a model at test-time to track a specific instance, we pretrain our generic tracker on a large variety of objects and efficiently update on the fly; Re3 simultaneously tracks and updates the appearance model with a single forward pass. This lightweight model is capable of tracking objects at 150 FPS, while attaining competitive results on challenging benchmarks. We also show that our method handles temporary occlusion better than other comparable trackers using experiments that directly measure performance on sequences with occlusion.

  Back
 
Keywords:
AI and DL Research, Intelligent Video Analytics and Smart Cities, Autonomous Machines, Computer Vision, Robotics & Autonomous Machines, GTC Silicon Valley 2018 - ID S8298
Streaming:
Share:
 
Training ImageNet In 15 Minutes With ChainerMN: A Scalable Distributed DL Framework
We''ll present a multi-node distributed deep learning framework called ChainerMN. Even though GPUs are continuously gaining more computation throughput, it is still very time-consuming to train state-of-the-art deep neural network models. For better ...Read More
We''ll present a multi-node distributed deep learning framework called ChainerMN. Even though GPUs are continuously gaining more computation throughput, it is still very time-consuming to train state-of-the-art deep neural network models. For better scalability and productivity, it is paramount to accelerate the training process by using multiple GPUs. To enable high-performance and flexible distributed training, ChainerMN was developed and built on top of Chainer. We''ll first introduce the basic approaches to distributed deep learning and then explain the design choice, basic usage, and implementation details of Chainer and ChainerMN. To demonstrate the scalability and efficiency of ChainerMN, we''ll discuss the remarkable results from training ResNet-50 classification model on ImageNet database using 1024 Tesla P100 GPUs and our in-house cluster, MN-1.    Back
 
Keywords:
AI and DL Research, NVIDIA Inception Program, Deep Learning and AI Frameworks, HPC and AI, GTC Silicon Valley 2018 - ID S8889
Streaming:
Download:
Share:
 
Towards Theory of AI's Mind
To effectively leverage the progress in Artificial Intelligence (AI) to make our lives more productive, it is important for humans and AI to work well together in a team. Traditionally, research has focused primarily on making AI more accurate, and ( ...Read More
To effectively leverage the progress in Artificial Intelligence (AI) to make our lives more productive, it is important for humans and AI to work well together in a team. Traditionally, research has focused primarily on making AI more accurate, and (to a lesser extent) on having it better understand human intentions, tendencies, beliefs, and contexts. The latter involves making AI more human-like and having it develop a theory of our minds. In this talk, I will argue that for human-AI teams to be effective, humans must also develop a Theory of AI''s Mind get to know its strengths, weaknesses, beliefs, and quirks. I will present some (very) initial results in the context of visual question answering and visual dialog where the AI agent is trained to answer natural language questions about images.  Back
 
Keywords:
AI and DL Research, Computer Vision, GTC Silicon Valley 2018 - ID S8560
Streaming:
Download:
Share:
 
Deep Learning for Computational Science
We''ll review our study of the use of artificial intelligence to augment various domains of computational science in order to improve time to solution for various HPC problems. We''ll discuss the current state-of-the-art approaches and performance ga ...Read More
We''ll review our study of the use of artificial intelligence to augment various domains of computational science in order to improve time to solution for various HPC problems. We''ll discuss the current state-of-the-art approaches and performance gains where applicable. We''ll also investigate current barriers to adoption and consider possible solutions.  Back
 
Keywords:
AI and DL Research, HPC and AI, GTC Silicon Valley 2018 - ID S8242
Streaming:
Download:
Share:
 
Deep Learning for Driver State Sensing
We''ll explore how deep learning approaches can be used for perceiving and interpreting the driver''s state and behavior during manual, semi-autonomous, and fully-autonomous driving. We''ll cover how convolutional, recurr ...Read More

We''ll explore how deep learning approaches can be used for perceiving and interpreting the driver''s state and behavior during manual, semi-autonomous, and fully-autonomous driving. We''ll cover how convolutional, recurrent, and generative neural networks can be used for applications of glance classification, face recognition, cognitive load estimation, emotion recognition, drowsiness detection, body pose estimation, natural language processing, and activity recognition in a mixture of audio and video data.

  Back
 
Keywords:
AI and DL Research, Autonomous Vehicles, Autonomous Driving, GTC Silicon Valley 2018 - ID S8626
Streaming:
Share:
 
Sparse-to-Dense: Depth Prediction from Sparse Depth Samples and a Single Image
Learn how to predict a dense depth image from a sparse set of depth measurements and a single RGB image. This approach can be applied to serve as a plug-in module in simultaneous localization and mapping to convert sparse maps to dense maps, and as a ...Read More
Learn how to predict a dense depth image from a sparse set of depth measurements and a single RGB image. This approach can be applied to serve as a plug-in module in simultaneous localization and mapping to convert sparse maps to dense maps, and as a super-resolution of LiDAR depth data. We''ll describe the performance of our prediction method, explain how to train the depth prediction network, and showcase examples of its applications. Codes and video demonstration are also publicly available. This session is for registrants who are already familiar with basic machine learning techniques.  Back
 
Keywords:
AI and DL Research, Computer Vision, GTC Silicon Valley 2018 - ID S8216
Streaming:
Share:
 
Additive Learning Framework for Self-Evolving AI
We''ll present a framework that can learn a compute-intensive deep neural networks (DNNs) task using multiple AI blocks and evolve better confidence by combining estimates. We''ll consider the example of establishing the identity of a user using spee ...Read More
We''ll present a framework that can learn a compute-intensive deep neural networks (DNNs) task using multiple AI blocks and evolve better confidence by combining estimates. We''ll consider the example of establishing the identity of a user using speech and image data. The system consists of two blocks - the AI block and Arbiter block. The AI block uses multiple DNNs (voice-based and image-based DNNs that generate a low confidence estimate initially). These AI blocks assist each other using Arbiter blocks and build confidence, improve accuracy, and learn salient features over time. Arbiter can store recent unacquainted data at run time in noisy and distorted environments and train the AI blocks periodically or on an on-demand basis. This concept could potentially improve the automatic speech recognition capabilities and allow detection of faces even when variable features of faces change with time. The GPU is the ideal choice as the task requires inferencing as well as training on the go.  Back
 
Keywords:
AI and DL Research, Intelligent Video Analytics and Smart Cities, Advanced AI Learning Techniques (incl. GANs and NTMs), GTC Silicon Valley 2018 - ID S8331
Streaming:
Download:
Share:
 
Attention GAN for Fine-Grained Language-to-Image Generation
We have long envisioned that machines one day can perform human-like perception, reasoning, and expression across multiple modalities including vision and language, which will augment and transform the ways humans communicate with each other and with ...Read More
We have long envisioned that machines one day can perform human-like perception, reasoning, and expression across multiple modalities including vision and language, which will augment and transform the ways humans communicate with each other and with the real world. With this vision, we''ll introduce the latest work of developing a deep attention GAN for fine-grained language-to-image synthesis. We''ll discuss the open problems behind the task that we''re thrilled to solve, including image and language understanding, joint reasoning across both modalities, and expressing abstract concepts into full imagination, which are of fundamental importance to reaching general intelligence.  Back
 
Keywords:
AI and DL Research, Advanced AI Learning Techniques (incl. GANs and NTMs), GTC Silicon Valley 2018 - ID S8867
Streaming:
Download:
Share:
 
Deep Gradient Compression: Reducing the Communication Bandwidth for Distributed Training
We find 99.9 percent of the gradient exchange in distributed SGD is redundant, and we propose deep gradient compression (DGC) to greatly reduce the communication bandwidth and improve the scalability of distributed training. To preserve accuracy duri ...Read More
We find 99.9 percent of the gradient exchange in distributed SGD is redundant, and we propose deep gradient compression (DGC) to greatly reduce the communication bandwidth and improve the scalability of distributed training. To preserve accuracy during this compression, DGC employs four methods: momentum correction, local gradient clipping, momentum factor masking, and warm-up training. We have applied DGC to image classification, speech recognition, and language modeling with multiple datasets including Cifar10, ImageNet, Penn Treebank, and Librispeech Corpus. In all these scenarios, DGC achieves a gradient compression ratio from 270x to 600x without losing accuracy, cutting the gradient size of ResNet-50 from 97MB to 0.35MB, and for DeepSpeech from 488MB to 0.74MB. DGC enables large-scale distributed training on inexpensive commodity 1Gbps Ethernet and facilitates distributed training on mobile.  Back
 
Keywords:
AI and DL Research, GTC Silicon Valley 2018 - ID S8607
Streaming:
Share:
 
Deep Learning for Recommender Systems
In this talk, we will survey how Deep Learning methods can be applied to personalization and recommendations. We will cover why standard Deep Learning approaches don''t perform better than typical collaborative filtering techniques. Then ...Read More

In this talk, we will survey how Deep Learning methods can be applied to personalization and recommendations. We will cover why standard Deep Learning approaches don''t perform better than typical collaborative filtering techniques. Then we will survey we will go over recently published research at the intersection of Deep Learning and recommender systems, looking at how they integrate new types of data, explore new models, or change the recommendation problem statement. We will also highlight some of the ways that neural networks are used at Netflix and how we can use GPUs to train recommender systems. Finally, we will highlight promising new directions in this space.

  Back
 
Keywords:
AI and DL Research, Consumer Engagement and Personalization, Deep Learning and AI, GTC Silicon Valley 2018 - ID S81011
Streaming:
Download:
Share:
 
Efficient Communication Library for Large-Scale Deep-Learning
We''ll talk about the challenges in a large-scale distributed, GPU-based deep learning, and propose an efficient communication algorithm to achieve state-of-the-art scalability. In detail, we''ll explain various ways to speed up GPU-based deep learni ...Read More
We''ll talk about the challenges in a large-scale distributed, GPU-based deep learning, and propose an efficient communication algorithm to achieve state-of-the-art scalability. In detail, we''ll explain various ways to speed up GPU-based deep learning, and motivate the large-scale deep-learning in the performance context. Then, we will state that efficient communication is a grand challenge in the large-scale deep-learning, especially with upcoming more powerful GPUs such as Volta architecture Tesla V100. We''ll present the technical details on a proposed communication algorithm along with the supporting data collected with more than 100 GPUs.  Back
 
Keywords:
AI and DL Research, Deep Learning and AI Frameworks, GTC Silicon Valley 2018 - ID S8479
Download:
Share:
 
Designing Human Centric Spaces with Holodeck and Machine Learning
The growth in density of housing in cities like London and New York has resulted in the higher demand for efficient smaller apartments. These designs challenge the use of space and function while trying to ensure the occupants have the perceptio ...Read More

The growth in density of housing in cities like London and New York has resulted in the higher demand for efficient smaller apartments. These designs challenge the use of space and function while trying to ensure the occupants have the perception of a larger space than provided. The process of designing these spaces has always been the responsibility and perception of a handful of designers using 2D and 3D static platforms as part of the overall building design and evaluation, typically constraint by a prescriptive program and functional requirement. A combination of human- and AI-based agents creating and testing these spaces through design and virtual immersive environments (NVIDIA Holodeck) will attempt to ensure the final results are efficient and best fit for human occupancy prior to construction.

  Back
 
Keywords:
AI and DL Research, Virtual Reality and Augmented Reality, GTC Silicon Valley 2018 - ID S8398
Streaming:
Download:
Share:
 
Learning Steering Bounds for Parallel Autonomy: Handling Ambiguity in End-to-End Driving
End-to-end learning is a powerful new strategy for training neural networks from perception to control. While such systems have been shown to perform well for reactionary control, the representation learned is not usable for higher level decision mak ...Read More
End-to-end learning is a powerful new strategy for training neural networks from perception to control. While such systems have been shown to perform well for reactionary control, the representation learned is not usable for higher level decision making, such as navigation. We''ll discuss the latest methodologies for training end-to-end systems for parallel autonomy, and demonstrate some of the shortcomings when such decision making capability is needed.  Back
 
Keywords:
AI and DL Research, Autonomous Vehicles, GTC Silicon Valley 2018 - ID S8605
Streaming:
Download:
Share:
 
Synthetic Data Generation for an All-in-One Driver Monitoring System
Driver monitoring systems are used to detect many driver attributes like gaze, head pose, eye openness, and other features pertaining to attention and assistance. We''ll present a synthetic method of generating data for training DNNs, which caters to ...Read More
Driver monitoring systems are used to detect many driver attributes like gaze, head pose, eye openness, and other features pertaining to attention and assistance. We''ll present a synthetic method of generating data for training DNNs, which caters to the above mentioned features of the subject. We use blender for generating synthetic images, powered by NVIDIA GPUs, which can be scaled to match training needs. Synthetic data generatation allows precise control over data points that are difficult to control in a real environment, like pupil dialation. This approach avoids noisy measurements and results in high accuracy without the need for a high-precision 3D sensor.  Back
 
Keywords:
AI and DL Research, Autonomous Vehicles, Advanced AI Learning Techniques (incl. GANs and NTMs), GTC Silicon Valley 2018 - ID S8324
Streaming:
Download:
Share:
 
Deep Learning For Intelligent Multi-Sensor Analytics
Go beyond working with a single sensor and enter the realm of Intelligent Multi-Sensor Analytics (IMSA). We''ll introduce concepts and methods for using deep learning with multi-sensor, or heterogenous, data. There are many resources and ...Read More

Go beyond working with a single sensor and enter the realm of Intelligent Multi-Sensor Analytics (IMSA). We''ll introduce concepts and methods for using deep learning with multi-sensor, or heterogenous, data. There are many resources and examples available for learning how to leverage deep learning with public imagery datasets. However, few resources exist to demonstrate how to combine and use these techniques to process multi-sensor data. As an example, we''ll introduce some basic methods for using deep learning to process radio frequency (RF) signals and make it a part of your intelligent video analytics solutions. We''ll also introduce methods for adapting existing deep learning frameworks for multiple sensor signal types (for example, RF, acoustic, and radar). We''ll share multiple use cases and examples for leveraging IMSA in smart city, telecommunications, and security applications.

  Back
 
Keywords:
AI and DL Research, Intelligent Video Analytics and Smart Cities, Autonomous Machines, GTC Silicon Valley 2018 - ID S8260
Streaming:
Download:
Share:
 
Accelerating Scientific Simulation with Generative Adversarial Networks
Many scientific and engineering fields increasingly rely on complex and time consuming computational simulation as part of the modern scientific workflow. In many applications, such as High Energy Particle Physics, Cosmology, Geophysics, and others, ...Read More
Many scientific and engineering fields increasingly rely on complex and time consuming computational simulation as part of the modern scientific workflow. In many applications, such as High Energy Particle Physics, Cosmology, Geophysics, and others, simulations are the computational bottleneck for producing and testing results. We introduce the usage of Generative Adversarial Networks (GAN) as a potential tool for speeding up expensive theoretical models and simulations in scientific and engineering applications, ushering in a new era of deep learning-powered scientific discovery. We will show that using a GAN-based High Energy Physics fast simulator on GPUs can provide speedups of up to 100,000x when compared to traditional simulation software, while retaining high levels of precision. Finally, we will discuss modeling and architectural considerations in this domain with the hope of directly empowering scientists and engineers in other fields to experiment with Generative Adversarial Networks in order to speed up simulation across scientific domains.  Back
 
Keywords:
AI and DL Research, Advanced AI Learning Techniques (incl. GANs and NTMs), HPC and AI, GTC Silicon Valley 2018 - ID S81001
Streaming:
Download:
Share:
 
Deep Reinforcement Learning for Real-World Robotic Manipulation
Deep reinforcement learning (deep RL) has emerged as a promising direction for autonomous acquisition of complex behaviors due to its ability to process complex sensory input and to acquire elaborate behavior skills, using general-purpose neural netw ...Read More
Deep reinforcement learning (deep RL) has emerged as a promising direction for autonomous acquisition of complex behaviors due to its ability to process complex sensory input and to acquire elaborate behavior skills, using general-purpose neural network representations. Since learning expressive function approximators requires large quantities of data, deep RL has been mostly applied to simulated domains, such as video games and simulated robotic locomotion and manipulation tasks, where the data collection can occur faster than real time and be trivially parallelized. We''ll address techniques that have been proposed to enable deep RL for real-world robotics, and discuss how the maximum-entropy principle can be leveraged to reduce the required amount of real-world interaction.  Back
 
Keywords:
AI and DL Research, GTC Silicon Valley 2018 - ID S8603
Streaming:
Download:
Share:
 
Generate Neural Network Automatically with High Accuracy and High Efficiency
Designing neural network architectures are critical for deep learning applications, but it is so complex and depends on AI experts. We''ll demonstrate how you can learn how to construct neural networks automatically without the human intervention. Th ...Read More
Designing neural network architectures are critical for deep learning applications, but it is so complex and depends on AI experts. We''ll demonstrate how you can learn how to construct neural networks automatically without the human intervention. There are two fundamental limiters to the performance of auto-generated neural networks: accuracy and efficiency, which is caused by searching overhead. We''ll also explore new techniques to make auto-generated neural network methods accurate and efficient, including: end-to-end technology to construct neural network within reinforcement learning, adaptive random search and bayesian optimization framework for different AI domains, such as computer vision, IoT acoustics, NLP and finance; using historical knowledge bases to reduce the searching overhead; and scheduling the execution of searching tasks over multiple NVIDIA GPUs to speed up the searching process. Also, we''ll give both the theoretical analysis and experiment results, which show significant improvement of accuracy and substantial reduction of searching time.  Back
 
Keywords:
AI and DL Research, GTC Silicon Valley 2018 - ID S8234
Streaming:
Download:
Share:
 
GPU Accelerated Sequence Learning for Action Recognition
We''ll introduce several attempts for modeling the long-term sequence dependence to help improve the action recognition performance. First, we''ll introduce a fused feature of deep and hand-crafted features to prove the complementation between them. ...Read More
We''ll introduce several attempts for modeling the long-term sequence dependence to help improve the action recognition performance. First, we''ll introduce a fused feature of deep and hand-crafted features to prove the complementation between them. We''ll also introduce an attempt of attention model to illustrate the effectiveness of attention mechanism on action recognition. We''ll then introduce shuttleNet, which is a biologically-inspired neural network. Finally, we''ll give some divergent experiments on action recognition to show the potential research direction.  Back
 
Keywords:
AI and DL Research, Computer Vision, Video and Image Processing, GTC Silicon Valley 2018 - ID S8229
Streaming:
Download:
Share:
 
The Future of the In-Car Experience
As the race to full autonomy accelerates, the in-cab transportation experience is also being redefined. Future vehicles will sense the passengers'' identities and activities, as well as their cognitive and emotional states, to adapt and ...Read More

As the race to full autonomy accelerates, the in-cab transportation experience is also being redefined. Future vehicles will sense the passengers'' identities and activities, as well as their cognitive and emotional states, to adapt and optimize their experience. AI capable of interpreting what we call "people analytics" captured through their facial and vocal expressions, and aspects of the context that surrounds them will power these advances. We''ll give an overview of our Emotion AI solution, and describe how we employ techniques like deep learning-based spatio-temporal modeling. By combining these techniques with a large-scale dataset, we can develop AI capable of redefining the in-cab experience.

  Back
 
Keywords:
AI and DL Research, NVIDIA Inception Program, Deep Learning and AI Frameworks, Autonomous Vehicles, GTC Silicon Valley 2018 - ID S8758
Streaming:
Download:
Share:
 
Smart City: Deep Learning Model for Car-Pedestrian Interaction
In this talk we will discuss the work Columbia University, in partnership with NYC government, is using deep learning and GPUs to develop smart city traffic management facilitating support for navigation/movement of multitude of vehicles (including a ...Read More
In this talk we will discuss the work Columbia University, in partnership with NYC government, is using deep learning and GPUs to develop smart city traffic management facilitating support for navigation/movement of multitude of vehicles (including autonomous cars) in dense urban environments with many pedestrians. We will describe our work in real-time tracking of cars and pedestrians, prediction of movement based on historical observations of the intersection, backed by ultra-low latency wireless communications and edge computing nodes.  Back
 
Keywords:
AI and DL Research, Intelligent Video Analytics and Smart Cities, Autonomous Vehicles, GTC Silicon Valley 2018 - ID S8201
Streaming:
Share:
 
Differentiable Tree Planning for Deep Reinforcement Learning
We''ll discuss recent research in deep reinforcement learning (RL), with a focus on the application of intuitions, from planning to neural network architectures for deep RL. Planning in complex visual environments has thus far been held back by the d ...Read More
We''ll discuss recent research in deep reinforcement learning (RL), with a focus on the application of intuitions, from planning to neural network architectures for deep RL. Planning in complex visual environments has thus far been held back by the difficulty of learning accurate predictive models. To address this, we embedded a model inside a differentiable, dynamically-constructed tree-planning architecture, so that we identify an effective model when used within that planner. We''ll share our work on developing these architectures, as well as our approaches to various technical obstacles associated with the efficient optimization of deep tree-structured models on GPU.  Back
 
Keywords:
AI and DL Research, Advanced AI Learning Techniques (incl. GANs and NTMs), GTC Silicon Valley 2018 - ID S8787
Streaming:
Download:
Share:
 
Training Deep AutoEncoders for Collaborative Filtering
This session will describe an approach to building personalized recommendations using (very) deep autoencoders. We will explore effects of different activation functions, network depth and novel algorithmic approaches. The model is trained end-to-end ...Read More
This session will describe an approach to building personalized recommendations using (very) deep autoencoders. We will explore effects of different activation functions, network depth and novel algorithmic approaches. The model is trained end-to-end without any layer-wise pre-training and our PyTorch-based code is publicly available.  Back
 
Keywords:
AI and DL Research, Consumer Engagement and Personalization, GTC Silicon Valley 2018 - ID S8212
Streaming:
Download:
Share:
 
GAN Fashion Photo Shoot: Garment to Model Images Using Conditional GANs
Learn how VUE.ai''s model generator uses conditional GANs to produce product-specific images suitable for replacing photographs in catalogs. We''ll present networks that generate images of fashion models wearing specific garments, using an image of t ...Read More
Learn how VUE.ai''s model generator uses conditional GANs to produce product-specific images suitable for replacing photographs in catalogs. We''ll present networks that generate images of fashion models wearing specific garments, using an image of the garment as a conditioning variable. Network architecture variants, training, and manipulation of latent variables to control attributes such as model pose, build, or skin color will be addressed.  Back
 
Keywords:
AI and DL Research, NVIDIA Inception Program, Advanced AI Learning Techniques (incl. GANs and NTMs), GTC Silicon Valley 2018 - ID S8776
Streaming:
Download:
Share:
 
Learning Affinity via Spatial Propagation Networks
We provide a unified framework on learning affinity in pure data-driven fashion using a linear propagation structure. This is a GPU and deep learning friendly pairwise learning module that does not require solving linear equation, iterative inference ...Read More
We provide a unified framework on learning affinity in pure data-driven fashion using a linear propagation structure. This is a GPU and deep learning friendly pairwise learning module that does not require solving linear equation, iterative inferences or manually defined kernels. Specifically, we develop a three-way connection for the linear propagation model, which formulates a sparse transformation matrix, where all elements can be the output from a deep CNN, but results in a dense affinity matrix that effectively models any task-specific pairwise similarity matrix. The spatial propagation network can be applied to many affinity-related tasks, such as image matting, segmentation and colorization, to name a few. Essentially, the model can learn semantically aware affinity relations for high-level vision tasks due to the powerful learning capability of the deep CNN. We validate the framework on the task of refinement for image segmentation boundaries. Experiments on face parsing and semantic segmentation tasks show that the spatial propagation network provides a general, effective, and efficient solution for generating high-quality segmentation results.  Back
 
Keywords:
AI and DL Research, Computer Vision, Video and Image Processing, GTC Silicon Valley 2018 - ID S8312
Streaming:
Download:
Share:
 
Scaling Deep Learning for Immersive User Interfaces
Deep learning creates advances following a virtuous recipe: model architecture search, creating large training datasets, and scaling computation. Baidu Research''s Silicon Valley AI Lab develops state-of-the-art conversational user interfaces followi ...Read More
Deep learning creates advances following a virtuous recipe: model architecture search, creating large training datasets, and scaling computation. Baidu Research''s Silicon Valley AI Lab develops state-of-the-art conversational user interfaces following this DL recipe. We research new model architectures and features for speech recognition (Deep Speech 3), speech generation (Deep Voice 3), and natural language processing. To deploy these models in impactful products, we want a deep understanding of how recipe components coordinate to drive accuracy improvements. Through large-scale empirical studies, we find intriguing results about how deep learning is likely to scale: As training set size increases, DL model generalization error and model sizes scale as particular power-law relationships. For a fixed dataset size, as model size grows, training time remains roughly constant -- larger models require fewer steps to converge to the same accuracy. These scaling relationships have significant implications on DL research, practice, and systems. They can assist model debugging, setting accuracy targets, and decisions about dataset growth and future computing system design.  Back
 
Keywords:
AI and DL Research, GTC Silicon Valley 2018 - ID S8899
Streaming:
Download:
Share:
 
Synthetic Facial Data for Training Deep Neural Networks
Training AI agents that can successfully generalize requires large amounts of diverse labeled training data. Collecting and labeling data is a significant cost in the development of AI applications, which, in some cases, may not even be feasib ...Read More
Training AI agents that can successfully generalize requires large amounts of diverse labeled training data. Collecting and labeling data is a significant cost in the development of AI applications, which, in some cases, may not even be feasible. We'll describe computer graphics facial models that we are developing to generate large labeled synthetic facial data for training deep neural networks. Facial analysis is central to many vision applications that involve human-computer interaction, including robotics, autonomous cars, rehabilitation, and extended usability. Generating and animating human faces with high realism is a well-studied problem in computer graphics; however, very few computer vision AI techniques take advantage of rendered facial data to augment or replace manually collected training data. We'll share key insights of how we successfully use synthetic facial data for training facial analysis classifiers. We'll also demonstrate many sub-tasks on which synthetic data helps to significantly improve accuracy and reduces the need for manual data collection.
 
  Back
 
Keywords:
AI and DL Research, Intelligent Video Analytics and Smart Cities, GTC Silicon Valley 2018 - ID S8794
Streaming:
Share:
 
De Novo Drug Design using Artificial Intelligence
We propose a novel computational strategy based on deep and reinforcement learning techniques for de-novo design of molecules with desired properties. This strategy integrates two deep neural networks generative and predictive to generate novel c ...Read More
We propose a novel computational strategy based on deep and reinforcement learning techniques for de-novo design of molecules with desired properties. This strategy integrates two deep neural networks generative and predictive to generate novel chemical structures with the desired properties. In the first phase of the method, generative and predictive models are separately trained with supervised learning algorithms. In the second phase, both models are jointly trained with reinforcement learning approach to bias newly generated chemical structures towards those with desired physical and biological properties. In this proof-of-concept study, we have employed this strategy to design chemical libraries biased toward compounds with either maximal, minimal, or specific range of physical properties, such as melting point and hydrophobicity, as well as to develop novel putative inhibitors of JAK2. This new approach can find a general use for generating targeted chemical libraries optimized for a single desired property or multiple properties.  Back
 
Keywords:
AI and DL Research, Computational Biology and Chemistry, GTC Silicon Valley 2018 - ID S8254
Streaming:
Download:
Share:
 
Towards Learning to Imagine Videos with Controlled Content
We discuss one of the first attempts to teach computers to imagine or generate videos with controlled content using deep learning generative modeling techniques. To this end, we assume visual information in a natural video can be decomposed into two ...Read More
We discuss one of the first attempts to teach computers to imagine or generate videos with controlled content using deep learning generative modeling techniques. To this end, we assume visual information in a natural video can be decomposed into two major components: content and motion. While content encodes the objects present in the video, motion encodes the object dynamics. Based on this prior, we propose the motion and content decomposed generative adversarial network (MoCoGAN) framework for video generation. The proposed framework generates a video clip by sequentially mapping random noise vectors to video frames. We divide a random noise vector into content and motion parts. By controlling these parts we generate both the content of the video and the action that is being performed. We perform quantitative and qualitative analysis on several video datasets, including artificial shape motion, facial expression, and tai-chi videos.  Back
 
Keywords:
AI and DL Research, Advanced AI Learning Techniques (incl. GANs and NTMs), Computer Vision, GTC Silicon Valley 2018 - ID S8477
Streaming:
Share:
 
Quick and Easy DL Workflow Proof of Concept
Spin up a deep learning (DL) proof-of-concept on a budget. We'll walk you through a DL workflow in the cloud leveraging DIGITS, then download a trained model, and run inference on a Jetson TX2. This session considers multiple options such as Nimbix, ...Read More
Spin up a deep learning (DL) proof-of-concept on a budget. We'll walk you through a DL workflow in the cloud leveraging DIGITS, then download a trained model, and run inference on a Jetson TX2. This session considers multiple options such as Nimbix, AMI, and NGC on Tesla P100, Tesla V100, and NVIDIA DGX-1 servers. This tutorial will be a combination of lecture, live demos, and detailed instructions.  Back
 
Keywords:
AI and DL Research, Accelerated Analytics, GTC Silicon Valley 2018 - ID S8286
Download:
Share:
 
Fully Context-Aware Video Prediction
We'll discuss the development of a novel model for video prediction and analysis -- the parallel multi-dimensional long short-term memory (PMD-LSTM). PMD-LSTM is a general model for learning from higher dimensional data such as images, videos, and b ...Read More
We'll discuss the development of a novel model for video prediction and analysis -- the parallel multi-dimensional long short-term memory (PMD-LSTM). PMD-LSTM is a general model for learning from higher dimensional data such as images, videos, and biomedical scans. It is an extension of the popular LSTM recurrent neural networks to higher dimensional data with a rearrangement of the recurrent connections to dramatically increase parallelism. This gives the network the ability to compactly model the effect of long-range context in each layer, unlike convolutional networks, which need several layers to cover a larger input context. We'll discuss the blind spot problem in recent work on video prediction, and show how PMD-LSTM based models are fully context-aware for each predicted pixel. These models outperform comparatively complex state-of-the-art approaches significantly in a variety of challenging video prediction scenarios such as car driving, human motion, and diverse human actions.  Back
 
Keywords:
AI and DL Research, NVIDIA Inception Program, Computer Vision, GTC Silicon Valley 2018 - ID S8713
Streaming:
Download:
Share:
 
SpaceNet: Accelerating Automated Feature Extraction for Satellite Imagery - Two years, Four Competitions in the Making
We'll present the results of the SpaceNet 2017-2018 Challenge, preview future SpaceNet Challenges, and how developers can generally access open labeled satellite image training data through SpaceNet on AWS. To date, three SpaceNet Challenges ha ...Read More
We'll present the results of the SpaceNet 2017-2018 Challenge, preview future SpaceNet Challenges, and how developers can generally access open labeled satellite image training data through SpaceNet on AWS. To date, three SpaceNet Challenges have been designed to apply computer vision techniques to satellite imagery which examine building footprint extraction, road network extraction, and off-nadir object detection. SpaceNet on AWS is an online repository of openly available satellite imagery, co-registered map data to train algorithms for developers and data scientists to access for research. This first-of-its-kind open innovation project for the geospatial industry launched in August 2016 as a collaboration between AWS, CosmiQ Works, DigitalGlobe, and NVIDIA. The SpaceNet Roads Challenge, launching in November, builds on labeled training datasets consisting of building footprints across Khartoum, Las Vegas, Paris, and Shanghai by providing over 8,000 km of mapped road networks. It uses a novel metric motivated by graph theory concepts that focused competitors on routing rather than just static road pixel identification.  Back
 
Keywords:
AI and DL Research, GIS, GTC Silicon Valley 2018 - ID S8553
Streaming:
Share:
 
Training Neural Networks with Mixed Precision: Theory and Practice
We'll cover the theory and practice for training DNNs with Tensor Cores, introduced for AI processing with the Volta GPU architecture. Tensor Cores provide up to 120 TFlops throughput, mixing operations on IEEE half- and single-precision floats. In ...Read More
We'll cover the theory and practice for training DNNs with Tensor Cores, introduced for AI processing with the Volta GPU architecture. Tensor Cores provide up to 120 TFlops throughput, mixing operations on IEEE half- and single-precision floats. In the theory portion of the talk, we'll review the half-precision format, values that arise in DNN computations, and techniques that maximize utilization of fp16 format by these values. Techniques include loss-scaling, master weights, and choosing the proper precision for a given operation. In the practice portion of this talk, we'll survey various models that have been trained in mixed precision, matching the accuracy of fp32 training sessions while using the same hyperparameters. Models include various architectures (feed forward, recurrent, generative) as well as cover diverse tasks (image, speech, and language processing). We'll also provide network design and training guidelines to maximize speed when using Tensor Cores.  Back
 
Keywords:
AI and DL Research, Algorithms and Numerical Techniques, GTC Silicon Valley 2018 - ID S8923
Streaming:
Download:
Share:
 
Deep Learning Applications in Text and Graphics at NVIDIA
At NVIDIA, we're busy applying deep learning to diverse problems, and this talk will give an overview of a few of these applications. We'll discuss our resume matching system, which helps match candidates to job openings at NVIDIA, as well as an op ...Read More
At NVIDIA, we're busy applying deep learning to diverse problems, and this talk will give an overview of a few of these applications. We'll discuss our resume matching system, which helps match candidates to job openings at NVIDIA, as well as an open-source sentiment analysis project trained on unsupervised text that is improving our marketing capabilities. We'll discuss a blind image quality metric that we're using to lower the cost of raytracing photorealistic graphics, and a generative model that we've built to create realistic graphics from simplistic sketches.  Back
 
Keywords:
AI and DL Research, GTC Silicon Valley 2018 - ID S8672
Streaming:
Share:
 
Unleashing the Imagination: Combining systems+software innovation with GPUs to create new capabilities
AI is one of the most rapidly-evolving areas of computer science today and datascientists are constantly pushing the boundaries of the possible -- wanting to explore new data types, new algorithms, and diverse and heterogenous models. In this talk we ...Read More
AI is one of the most rapidly-evolving areas of computer science today and datascientists are constantly pushing the boundaries of the possible -- wanting to explore new data types, new algorithms, and diverse and heterogenous models. In this talk we'll explore two key productivity factors for datascience -- first, speed and the ability to explore many models and sets of data quickly; and second, ability to explore broad types of models, incorporating both machine learning and deep learning. We will talk about results of 40x and 50x productivity through system+software co-design and novel algorithms which leverage Power Systems and GPUs for both deep learning and key areas of classical machine learning. System+software co-design and co-optimization can result in dramatic efficiency improvements, enable creation of large models, exploration of large datasets, and realize productivity gains for datascientists, freeing them up to focus on the fundamental science of deep and machine learning -- gaining accuracy, functionality, and generalizability of their models.  Back
 
Keywords:
AI and DL Research, GTC Silicon Valley 2018 - ID S81025
Streaming:
Share:
 
Overcoming Missing Modalities in Remote Sensing
Recent advances in earth observation are opening up a new exciting area for exploration of satellite image data. We'll teach you how to analyse this new data source with deep neural networks. Focusing on emergency response, you will learn how to app ...Read More
Recent advances in earth observation are opening up a new exciting area for exploration of satellite image data. We'll teach you how to analyse this new data source with deep neural networks. Focusing on emergency response, you will learn how to apply deep neural networks for semantic segmentation on satellite imagery. We will specifically focus on multimodal segmentation and the challenge of overcoming missing modality information during inference time. It is assumed that registrants are already familiar with fundamentals of deep neural networks.  Back
 
Keywords:
AI and DL Research, Advanced AI Learning Techniques (incl. GANs and NTMs), GTC Silicon Valley 2018 - ID S8596
Streaming:
Download:
Share:
 
Adaptive Ray Tracing Rendering Powered by Deep Learning
This session will present a proof of concept where a deep neural network was trained with pairs of Iray ray traced images (one arbitrary ray tracing iteration number and one fully converged image) and theirs structural similarity index (SSIM). Origin ...Read More
This session will present a proof of concept where a deep neural network was trained with pairs of Iray ray traced images (one arbitrary ray tracing iteration number and one fully converged image) and theirs structural similarity index (SSIM). Originally thought as a method for measuring the similarity between two images, SSIM index can also be viewed as a quality measure versus a reference image or, in our case, as a ray tracing rendering progress. The DNN can now from any render iteration of arbitrary scene infer a rendering progress estimator but also provides heat map pictures of the scenes that can be used for adaptive rendering, focusing ray tracing engine power on appropriate zones.  Back
 
Keywords:
AI and DL Research, Graphics and AI, Rendering and Ray Tracing, GTC Silicon Valley 2018 - ID S8788
Streaming:
Download:
Share:
 
SSD++ Boosting Performance of Single-Shot MultiBox Detection Using Convolution Autoencoders
We'll showcase how you can apply a wealth of unlabeled image data to significantly improve accuracy and speed of single-shot object-detection (SSD) techniques. Our approach, SSD++, advances the state-of-the-art of single shot multibox-based object d ...Read More
We'll showcase how you can apply a wealth of unlabeled image data to significantly improve accuracy and speed of single-shot object-detection (SSD) techniques. Our approach, SSD++, advances the state-of-the-art of single shot multibox-based object detectors (such as SSD, YOLO) by employing a novel combination of convolution-deconvolution networks to learn robust feature maps, thus making use of unlabeled dataset, and the fresh approach to have confluence of convolution and deconvolution features to combine generic as well as semantically rich feature maps. As a result, SSD++ drastically reduces the requirement of labeled datasets, works on low-end GPUs, identifies small as well as large objects with high fidelity, and speeds up inference process by decreasing the requirement of default boxes. SSD++ achieves state-of-the-art results on PASCAL VOC and MS COCO datasets. Through ablation study, we'll explain the effectiveness of different components of our architecture that help us achieve improved accuracy on the above datasets. We'll further show a case study of SSD++ to identify shoppable objects in fashion, home decor, and food industry from images in the wild.  Back
 
Keywords:
AI and DL Research, NVIDIA Inception Program, Computer Vision, Video and Image Processing, GTC Silicon Valley 2018 - ID S8159
Streaming:
Share:
 
Deep Learning for Dialogue Systems
Learn how to apply deep learning technologies for building robust and scalable dialogue systems with deeper understanding of the classic pipelines as well as detailed knowledge on the benchmark of the models of the prior work. We'll start with an ov ...Read More
Learn how to apply deep learning technologies for building robust and scalable dialogue systems with deeper understanding of the classic pipelines as well as detailed knowledge on the benchmark of the models of the prior work. We'll start with an overview of the dialogue research and allow the audience to dive deep into the state-of-the-art work about neural-based language understanding, dialogue management, and language generation towards end-to-end neural dialogue systems.  Back
 
Keywords:
AI and DL Research, Speech and Language Processing, GTC Silicon Valley 2018 - ID S8542
Streaming:
Download:
Share:
 
IamNN: Iterative and Adaptive Mobile Neural Network for Efficient Image Classification
Deep residual networks (ResNets) made a recent breakthrough in deep learning. The core idea of ResNets is to have shortcut connections between layers that allow the network to be much deeper while still being easy to optimize avoiding vanishing ...Read More

Deep residual networks (ResNets) made a recent breakthrough in deep learning. The core idea of ResNets is to have shortcut connections between layers that allow the network to be much deeper while still being easy to optimize avoiding vanishing gradients. These shortcut connections have interesting properties that make ResNets behave differently from other typical network architectures. In this talk we will use these properties to design a network based on a ResNet but with parameter sharing and adaptive computation time, we call it IamNN. The resulting network is much smaller than the original network and can adapt the computational cost to the complexity of the input image. During this talk we will provide an overview of ways to design compact networks, give an overview of ResNets properties and discuss how they can be used to design compact dense network with only 5M parameters for ImageNet classification.

  Back
 
Keywords:
AI and DL Research, GTC Silicon Valley 2018 - ID S8456
Streaming:
Download:
Share:
 
Multimodal Memory Modelling for Video Captioning
This talk presents a novel framework named multimodal memory model for video captioning, which builds a visual and textual shared memory to model the long-term visual-textual dependency and further guide visual attention on described visual targets t ...Read More
This talk presents a novel framework named multimodal memory model for video captioning, which builds a visual and textual shared memory to model the long-term visual-textual dependency and further guide visual attention on described visual targets to solve visual-textual alignments. Video captioning which automatically translates video clips into natural language sentences is a very important task in computer vision. By virtue of recent deep learning technologies, video captioning has made great progress. However, learning an effective mapping from the visual sequence space to the language space is still a challenging problem due to the long-term multimodal dependency modelling and semantic misalignment. Inspired by the facts that memory modelling poses potential advantages to long-term sequential problems and working memory is the key factor of visual attention, the proposed model attaches an external memory to store and retrieve both visual and textual contents by interacting with video and sentence with multiple read and write operations.  Back
 
Keywords:
AI and DL Research, Computer Vision, Video and Image Processing, GTC Silicon Valley 2018 - ID S8311
Streaming:
Download:
Share:
 
Getting Started with Tensorflow on GPUs
Want to get started using TensorFlow together with GPUs? Then come to this session, where we will cover the TensorFlow APIs you should use to define and train your models, and the best practices for distributing the training workloads to multipl ...Read More

Want to get started using TensorFlow together with GPUs? Then come to this session, where we will cover the TensorFlow APIs you should use to define and train your models, and the best practices for distributing the training workloads to multiple GPUs. We will also look at the underlying reasons why are GPUs are so great to use for Machine Learning workloads?

  Back
 
Keywords:
AI and DL Research, Deep Learning and AI, Developer Talk, GTC Silicon Valley 2018 - ID S8946
Streaming:
Download:
Share:
 
Sim2Real Visual Robotic Servoing for Navigation and Manipulation via Deep Reinforcement Learning
Humans are remarkably proficient at controlling their limbs and tools from a wide range of viewpoints, diverse environments and in the presence of distractors. In robotics, this ability is referred to as visual servoing. Standard visual servoing appr ...Read More
Humans are remarkably proficient at controlling their limbs and tools from a wide range of viewpoints, diverse environments and in the presence of distractors. In robotics, this ability is referred to as visual servoing. Standard visual servoing approaches have limited generalization as they typically rely on manually designed features and calibrated camera. We exhibit generalizable visual servoing in the context of robotic manipulation and navigation tasks learned through visual feedback and by deep reinforcement learning (RL) without needing any calibrated setup. By highly randomizing our simulator, we train policies that generalize to novel environments and also to the challenging real world scenarios. Our domain randomization technique addresses the high sample complexity of deep RL, avoids the dangers of trial-and-error and also provides us with the liberty to learn recurrent vision-based policies for highly diverse tasks where capturing sufficient real robot data is impractical. An example of such scenario is learning view-invariant robotic policies which leads into learning physical embodiment and self-calibration purely through visual feedback.  Back
 
Keywords:
AI and DL Research, IoT, Robotics & Drones, Robotics & Autonomous Machines, GTC Silicon Valley 2018 - ID S8955
Streaming:
Share:
 
Investigating Data Augmentation Strategies for Advancing Deep Learning Training
We saw the huge success of the deep learning paradigm and the superhuman capability in numerous benchmarks in image, video, audio, or text. However, it poses huge challenges as adopting the methods in industrial applications (mainly due to the lack o ...Read More
We saw the huge success of the deep learning paradigm and the superhuman capability in numerous benchmarks in image, video, audio, or text. However, it poses huge challenges as adopting the methods in industrial applications (mainly due to the lack of quality tracking data) as the neural networks consume enormous parameters and require relatively huge quality training data. We'll aim for investigating the "data augmentation" strategies increasing quality training data for robust inference across different learning problems mainly in image, video, 3D, and IoT data streams. We'll first quantify the importance of training data for deep neural networks then review numerous strategies, such as crawling from the web, utilizing generative models, 3D computer graphics, augmented reality, engagement in social media, gaming, etc. We'll compare the effectiveness among the diverse strategies. As generally taking the data from other domains, we also need to deal with the cross-domain learning problem. We'll provide detailed insights from our recent work published in top conferences (e.g., CVPR, ICCV, AAAI, etc.) and those cases in industrial applications.  Back
 
Keywords:
AI and DL Research, Advanced AI Learning Techniques (incl. GANs and NTMs), GTC Silicon Valley 2018 - ID S8391
Streaming:
Download:
Share:
 
Tensor Layers for Compression of Deep Learning Networks
We'll review recent efforts to compress fully connected layers in machine learning via tensor networks, including the Tensor Train format, the Tensor Contraction Layer, the Tensor Regression Layer, and a Tensor Ring decomposition. These decompositio ...Read More
We'll review recent efforts to compress fully connected layers in machine learning via tensor networks, including the Tensor Train format, the Tensor Contraction Layer, the Tensor Regression Layer, and a Tensor Ring decomposition. These decompositions, in supplementing or replacing fully connected layers, are shown to dramatically reduce the number of parameters required by the network without resorting to sparsity and without loss in error. We've shown 55-80 percent compression of the entire network with less than one percent loss in accuracy. These Tensor layers can be used in end-to-end training, fine-tuning, and transfer-learning by initializing the decomposition with a pre-trained fully connected layer. Furthermore, because the forward and backward passes of the network rely on dense Tensor contractions, we show that these methods retain high computational intensity and can be efficiently evaluated on GPUs.  Back
 
Keywords:
AI and DL Research, Algorithms and Numerical Techniques, HPC and AI, GTC Silicon Valley 2018 - ID S8807
Streaming:
Download:
Share:
 
Accelerating Cancer Research with Deep Learning
The Department of Energy (DOE) entered into a partnership with the National Cancer Institute (NCI) of the National Institutes of Health (NIH) to accelerate cancer research. This "Cancer Moonshot" aims to tackle three main objectives: better ...Read More
The Department of Energy (DOE) entered into a partnership with the National Cancer Institute (NCI) of the National Institutes of Health (NIH) to accelerate cancer research. This "Cancer Moonshot" aims to tackle three main objectives: better understand the mechanisms of cancer, use large amounts of diverse medical data for predictive models, and enable precision medicine by providing guidance for treatment to individual patients. Leveraging the compute expertise of DOE in high performance computing (HPC) and new methods for deep learning in artificial intelligence, this HPC+AI approach aims to create a single scalable deep neural network code called CANDLE (CANcer Distributed Learning Environment) that will be used to address all three challenges. This talk aims to give an overview of the project and highlight how GPU accelerated systems in the DOE ecosystem, Summit and Sierra, have contributed to the project.  Back
 
Keywords:
AI and DL Research, HPC and AI, Medical Imaging and Radiology, GTC Silicon Valley 2018 - ID S81033
Streaming:
Share:
 
Multi-Resolution 3D-Convolutional Neural Network for Object Recognition
Voxelized representation of 3D objects is commonly used for training 3D-Convolutional Neural Networks for object detection and classification. However, high-resolution voxelization of CAD models are memory intensive and hence, it is not possible to l ...Read More
Voxelized representation of 3D objects is commonly used for training 3D-Convolutional Neural Networks for object detection and classification. However, high-resolution voxelization of CAD models are memory intensive and hence, it is not possible to load multiple models in the GPU for training. We have developed a GPU-accelerated voxelization technique that generates multi-level voxel grids of 3D objects. Instead of creating a single high-resolution voxel grid for the whole object, this technique generates selective region-based high-resolution voxel grids to represent detailed features in the object. We have also developed a multi-resolution 3D-Convolutional Neural Network that uses this hybrid voxelization for accurate object recognition and classification.  Back
 
Keywords:
AI and DL Research, Industrial Inspection, Computer Vision, GTC Silicon Valley 2018 - ID S8389
Streaming:
Download:
Share:
 
High-Resolution Image Synthesis and Semantic Manipulation with Conditional GANs
We'll present a new method for synthesizing high-resolution photo-realistic images from semantic label maps using conditional generative adversarial networks. Conditional GANs have enabled a variety of applications, but the results are often limited ...Read More
We'll present a new method for synthesizing high-resolution photo-realistic images from semantic label maps using conditional generative adversarial networks. Conditional GANs have enabled a variety of applications, but the results are often limited to low-res and still far from realistic. We'll show that we're capable of generating 2048x1024 visually appealing results with a novel adversarial loss, as well as new multi-scale generator and discriminator architectures. Furthermore, we extend our framework to interactive visual manipulation with two additional features. First, we incorporate object instance segmentation information, which enables object manipulations such as removing/adding objects and changing the object category. Second, we propose a method to generate diverse results given the same input, allowing users to edit the object appearance interactively. Human opinion studies demonstrate that our method significantly outperforms existing methods, advancing both the quality and the resolution of deep image synthesis and editing.  Back
 
Keywords:
AI and DL Research, Graphics and AI, GTC Silicon Valley 2018 - ID S8918
Streaming:
Share:
 
GPU Performance Testing and PowerAI on IBM Cloud (Presented by IBM Cloud)
In this session, you will learn about the latest IBM PowerAI solution, IBM Cloud GPU offerings and see a price-performance comparison, with supporting data, on the number of CPUs required to optimize GPU performance. We've also aggregated extensive ...Read More
In this session, you will learn about the latest IBM PowerAI solution, IBM Cloud GPU offerings and see a price-performance comparison, with supporting data, on the number of CPUs required to optimize GPU performance. We've also aggregated extensive test data to determine general best practices such as half-precision deep learning advantages on the Tesla V100 and the implications of neural-network model variable distribution and gradient aggregation techniques on your performance results. Join us to see why NVIDIA GPUs on IBM Cloud offer superior results.  Back
 
Keywords:
AI and DL Research, Accelerated Analytics, GTC Silicon Valley 2018 - ID S81013
Streaming:
Download:
Share: