SEARCH SESSIONS

Search All
 
Refine Results:
 
Year(s)

SOCIAL MEDIA

EMAIL SUBSCRIPTION

 
 

GTC On-Demand

Application Design & Porting Techniques
Presentation
Media
The NVIDIA Application Lab at Julich, established by JSC and NVIDIA in June 2012, aims on enabling scientific applications for GPU-based architectures. Selected applications and their performance characteristics will be presented. Strategies for ...Read More

The NVIDIA Application Lab at Julich, established by JSC and NVIDIA in June 2012, aims on enabling scientific applications for GPU-based architectures. Selected applications and their performance characteristics will be presented. Strategies for multi-GPU parallelizations (necessary to meet computing demands) will be discussed.

  Back
 
Keywords:
Application Design & Porting Techniques, Supercomputing 2012 - ID SC2007
Download:
Share:
Astronomy and Astrophysics
Presentation
Media
"Big Data" Astronomical Data Analysis and Visualization
I will present a high-performance; graphics processing unit (GPU)-based framework for the efficient analysis and visualization of ``big data'' astronomical data cubes. Using a cluster of 96 GPUs, we demonstrate for a 0.5 TB image: volume ...Read More

I will present a high-performance; graphics processing unit (GPU)-based framework for the efficient analysis and visualization of ``big data'' astronomical data cubes. Using a cluster of 96 GPUs, we demonstrate for a 0.5 TB image: volume rendering at 10 fps; computation of basic statistics in 1.7 s; and evaluation of the median in 45s. The framework is one of the first solutions to the image analysis and visualization requirements of next-generation telescopes, including the forthcoming SKA pathfinder telescopes.

  Back
 
Keywords:
Astronomy and Astrophysics, Supercomputing 2012 - ID SB001
Download:
Share:
Climate, Weather, Ocean Modeling
Presentation
Media
Running the FIM and NIM Weather Models on GPUs
Two U.S. global-scale weather models, developed at NOAA, are running on GPUs. The FIM runs at 15 KM resolution and is expected to be run by the U.S. National Weather Service in the next year. The NIM is a next-generation forecast model designed ...Read More

Two U.S. global-scale weather models, developed at NOAA, are running on GPUs. The FIM runs at 15 KM resolution and is expected to be run by the U.S. National Weather Service in the next year. The NIM is a next-generation forecast model designed to run at 4KM resolution. This presentation will give an update on our efforts to parallelize and run these models on GPUs.

  Back
 
Keywords:
Climate, Weather, Ocean Modeling, Supercomputing 2012 - ID SC2018
Download:
Share:
 
Hybrid CPU-GPU Solutions for Weather and Cloud Resolving Climate Simulations
Reliable weather prediction for the Alpine region and cloud resolving climate modeling require simulations that run at 1-2 km resolution. Additionally, since the largest possible ensembles are needed, high fidelity models have to run on the most ...Read More

Reliable weather prediction for the Alpine region and cloud resolving climate modeling require simulations that run at 1-2 km resolution. Additionally, since the largest possible ensembles are needed, high fidelity models have to run on the most economical resource in a given time to solution. In this presentation we will give an update on the refactoring of COSMO, a widely used production code in academia as well as seven European weather services, and discuss the performance experience on hybrid CPU-GPU systems.

  Back
 
Keywords:
Climate, Weather, Ocean Modeling, Supercomputing 2012 - ID SC2036
Download:
Share:
Cloud Visualization
Presentation
Media
Accelerating Compute-Intensive Processing with Hybrid GPU Parallelization and Cloud Computing
In this presentation, Altair will discuss how innovative hybrid parallelization using multiple GPUs and MPI dramatically reduces runtime for certain classes of compute-intensive workloads. Offloading intensive computations on the GPU and using h ...Read More

In this presentation, Altair will discuss how innovative hybrid parallelization using multiple GPUs and MPI dramatically reduces runtime for certain classes of compute-intensive workloads. Offloading intensive computations on the GPU and using heterogeneous computing with optimized workload management improves performance; users also benefit from simplified, accelerated access to compute resources via cloud portals.

  Back
 
Keywords:
Cloud Visualization, HPC and Supercomputing, Supercomputing 2012 - ID SC2026
Download:
Share:
Clusters & GPU Management
Presentation
Media
Tesla Cluster Monitoring and Management APIs
Learn more about cluster management and monitoring of NVIDIA GPUs. This includes a detailed description of the NVIDIA Management Library (NVML) and user-facing third party software. Additionally, the nvidia-healthmon GPU health check tool will b ...Read More

Learn more about cluster management and monitoring of NVIDIA GPUs. This includes a detailed description of the NVIDIA Management Library (NVML) and user-facing third party software. Additionally, the nvidia-healthmon GPU health check tool will be covered.

  Back
 
Keywords:
Clusters & GPU Management, Supercomputing 2012 - ID SC135
Share:
Computational Fluid Dynamics
Presentation
Media
Accelerated ANSYS Fluent: Algebraic Multigrid on a GPU
NVIDIAs NVAMG library is a sophisticated suite of multi-level linear solvers. I will present an overview of our approach to parallelizing all phases of algebraic multigrid, including hierarchy construction and ILU factorization and solve. I will ...Read More

NVIDIAs NVAMG library is a sophisticated suite of multi-level linear solvers. I will present an overview of our approach to parallelizing all phases of algebraic multigrid, including hierarchy construction and ILU factorization and solve. I will also describe how NVAMG provides GPU acceleration to the coupled incompressible solver in ANSYS Fluent 14.5.

  Back
 
Keywords:
Computational Fluid Dynamics, Supercomputing 2012 - ID SC2030
Download:
Share:
Computer Aided Engineering
Presentation
Media
Improving Engineering Productivity with HPC and GPU-Accelerated Simulation
High Performance Computing has been a mainstay of increased productivity for years now. But recently, GPUs have enabled another level of performance without the significant purchase and power consumption required by additional nodes. ANSYS, Inc. ...Read More

High Performance Computing has been a mainstay of increased productivity for years now. But recently, GPUs have enabled another level of performance without the significant purchase and power consumption required by additional nodes. ANSYS, Inc. continues to develop customer focused HPC solutions incorporating the latest hardware technologies including NVIDIA GPUs.

  Back
 
Keywords:
Computer Aided Engineering, Simulation for Autonomous Vehicles, Supercomputing 2012 - ID SC2027
Download:
Share:
Developer - Programming Languages
Presentation
Media
Using CAPS Compilers on NVIDIA Kepler and CARMA Systems
CAPS compilers provide directive based programming for the Kepler and CARMA systems. They support OpenACC and OpenHMPP directive styles. CAPS compilers are based on a source-to-source specific technology that allows to leverage accelerator's ...Read More

CAPS compilers provide directive based programming for the Kepler and CARMA systems. They support OpenACC and OpenHMPP directive styles. CAPS compilers are based on a source-to-source specific technology that allows to leverage accelerator's native compilers as well as host ones. This talk is focused on achieving code portability with other accelerator technologies.

  Back
 
Keywords:
Developer - Programming Languages, Supercomputing 2012 - ID SC2010
Download:
Share:
 
CUDA: Past, Present and Future
The GPU evolved from its humble beginnings as a VGA Accelerator to become a massively parallel general processor for heterogeneous computing systems. Once the opportunity became obvious, the challenge was how to best develop a general purpose pr ...Read More

The GPU evolved from its humble beginnings as a VGA Accelerator to become a massively parallel general processor for heterogeneous computing systems. Once the opportunity became obvious, the challenge was how to best develop a general purpose programming model to preserve the GPUs architectural advantage. Learn how CUDA as a parallel computing platform and programming model came to be, how its being leveraged in a wide range of fields, and get an exciting preview of where its going.

  Back
 
Keywords:
Developer - Programming Languages, Supercomputing 2012 - ID SC2011
Download:
Share:
 
Bringing GPU Computing to the Masses with OpenACC
The OpenACC standard has emerged as a solution for GPU computing in projects where CUDA programming resources are not available, or where code maintenance issues prevent its use. We have been teaching OpenACC to scientists and others with great ...Read More

The OpenACC standard has emerged as a solution for GPU computing in projects where CUDA programming resources are not available, or where code maintenance issues prevent its use. We have been teaching OpenACC to scientists and others with great success and will discuss this approach.

  Back
 
Keywords:
Developer - Programming Languages, HPC and AI, Supercomputing 2012 - ID SC2013
Download:
Share:
 
Productive Performance on the Cray XK System Using OpenACC Compilers and Tools
The Cray XK high level parallel programming environment was developed to help the widespread adoption of GPUs in HPC. Ease of use is possible with OpenACC compilers, making it feasible for users to write applications in Fortran, C, or C++, and t ...Read More

The Cray XK high level parallel programming environment was developed to help the widespread adoption of GPUs in HPC. Ease of use is possible with OpenACC compilers, making it feasible for users to write applications in Fortran, C, or C++, and tools and libraries to help users port, debug, and optimize for hyrid systems.

  Back
 
Keywords:
Developer - Programming Languages, Supercomputing 2012 - ID SC2016
Download:
Share:
 
Compiling Parallel Languages with the NVIDIA Compiler SDK
NVIDIA's CUDA C/C++ Compiler (NVCC) is based on the widely used LLVM open source compiler infrastructure. This open foundation enables developers to create or extend programming languages with support for GPU acceleration using the CUDA Comp ...Read More

NVIDIA's CUDA C/C++ Compiler (NVCC) is based on the widely used LLVM open source compiler infrastructure. This open foundation enables developers to create or extend programming languages with support for GPU acceleration using the CUDA Compiler SDK. In this talk you will learn how to use the NVIDIA Compiler SDK to generate high-performance parallel code for NVIDIA GPUs.

  Back
 
Keywords:
Developer - Programming Languages, Supercomputing 2012 - ID SC2019
Streaming:
Download:
Share:
 
New Features in CUDA 5
The performance and efficiency of CUDA, combined with a thriving ecosystem of programming languages, libraries, tools, training, and services, have helped make GPU computing a leading HPC technology. In this talk you will learn about powerful ne ...Read More

The performance and efficiency of CUDA, combined with a thriving ecosystem of programming languages, libraries, tools, training, and services, have helped make GPU computing a leading HPC technology. In this talk you will learn about powerful new features of CUDA 5 and the Kepler GPU architecture, including CUDA Dynamic Parallelism, CUDA device code linking, and the new NSight Eclipse Edition.

  Back
 
Keywords:
Developer - Programming Languages, Supercomputing 2012 - ID SC2020
Download:
Share:
 
Introduction to CUDA C/C++
Learn how to access the massively parallel processing power of NVIDIA GPUs using CUDA C and C++. Well start with a simple Hello Parallelism! program and progress on to something a little more complicated. You will see what actually happens when ...Read More

Learn how to access the massively parallel processing power of NVIDIA GPUs using CUDA C and C++. Well start with a simple Hello Parallelism! program and progress on to something a little more complicated. You will see what actually happens when you compile & run and how to add GPU+CPU hybrid computing concepts to accelerate your applications.

  Back
 
Keywords:
Developer - Programming Languages, Supercomputing 2012 - ID SC2021
Download:
Share:
 
Compiling Python to the GPU with Numba
GPUs can offer orders of magnitude speed-ups for certain calculations, but programming the GPU remains difficult. Using NVIDIA's new support of LLVM, Continuum Analytics has built an array-oriented compiler for Python called Numba that can t ...Read More

GPUs can offer orders of magnitude speed-ups for certain calculations, but programming the GPU remains difficult. Using NVIDIA's new support of LLVM, Continuum Analytics has built an array-oriented compiler for Python called Numba that can target the GPU. In this talk, I will demonstrate how Numba makes programming the GPU as easy as a one-line change to working Python code.

  Back
 
Keywords:
Developer - Programming Languages, Supercomputing 2012 - ID SC2037
Download:
Share:
Developer - Tools & Libraries
Presentation
Media
CUDA Development Using NVIDIA Nsight, Eclipse Edition
NVIDIA Nsight, Eclipse Edition for Linux and Mac is an all-in-one development environment that lets you develop, debug and optimize CUDA code in an integrated UI environment. This talk provides a detail usage walk-through of the fully CUDA aware ...Read More

NVIDIA Nsight, Eclipse Edition for Linux and Mac is an all-in-one development environment that lets you develop, debug and optimize CUDA code in an integrated UI environment. This talk provides a detail usage walk-through of the fully CUDA aware source editor, build integration of the CUDA tool chain, graphical debugger for both CPU and GPU, and graphical profiler to enable performance optimization. If youve been waiting for a CUDA IDE on Linux and Mac then this talk is for you.

  Back
 
Keywords:
Developer - Tools & Libraries, Supercomputing 2012 - ID SC2006
Download:
Share:
 
MAGMA - a New Generation of Linear Algebra Libraries for GPU and Multicore Architectures
This talk will highlight the latest accomplishments in the Matrix Algebra on GPU and Multicore Architectures (MAGMA) project. We use a hybridization methodology that is built on representing linear algebra algorithms as collections of tasks and ...Read More

This talk will highlight the latest accomplishments in the Matrix Algebra on GPU and Multicore Architectures (MAGMA) project. We use a hybridization methodology that is built on representing linear algebra algorithms as collections of tasks and data dependencies, as well as properly scheduling the tasks' execution over the available multicore and GPU hardware components. This methodology is applied in MAGMA to develop high-performance fundamental linear algebra routines, such as the one-sided dense matrix factorizations (LU, QR, and Cholesky) and linear solvers, two-sided dense matrix factorizations (bidiagonal, tridiagonal, and Hessenberg reductions) for singular and eigenvalue problems, in addition to iterative linear and eigenvalue solvers. MAGMA is designed to be similar to LAPACK in functionality, data storage, and interface, in order to allow scientists to effortlessly port any of their LAPACK-relying software components to take advantage of the new architectures.

  Back
 
Keywords:
Developer - Tools & Libraries, Supercomputing 2012 - ID SC2012
Download:
Share:
 
CUDA-Accelerated Libraries
This presentation will be an overview of several libraries in the CUDA SDK and other third-party libraries including cuBLAS, cuRAND, NPP and Thrust. Using these libraries can often significantly shorten the development time of a GPU project whil ...Read More

This presentation will be an overview of several libraries in the CUDA SDK and other third-party libraries including cuBLAS, cuRAND, NPP and Thrust. Using these libraries can often significantly shorten the development time of a GPU project while leading to high-performance, high-quality software. We will discuss common use cases and the strengths of individual libraries and also provide guidance for selecting the best library for your project.

  Back
 
Keywords:
Developer - Tools & Libraries, Supercomputing 2012 - ID SC2015
Download:
Share:
 
High-Productivity Development with the Thrust Parallel Algorithms Library
Thrust is a parallel algorithms library which resembles the C++ Standard Template Library (STL). Thrust's high-level interface greatly enhances developer productivity while enabling performance portability between GPUs and multicore CPUs. In ...Read More

Thrust is a parallel algorithms library which resembles the C++ Standard Template Library (STL). Thrust's high-level interface greatly enhances developer productivity while enabling performance portability between GPUs and multicore CPUs. Interoperability with established technologies (such as CUDA, TBB and OpenMP) facilitates integration with existing software. In this talk we'll walk through the library's main features and explain how developers can build high-performance applications rapidly with Thrust.

  Back
 
Keywords:
Developer - Tools & Libraries, Supercomputing 2012 - ID SC2035
Download:
Share:
HPC and AI
Presentation
Media
Inside the Kepler Architecture
This presentation looks into the features of NVIDIA's latest Kepler GPU architecture. Join us as one of CUDA's language architects explains what's new, why it's exciting, and demonstrates the power of Kepler GPU accelerators with ...Read More

This presentation looks into the features of NVIDIA's latest Kepler GPU architecture. Join us as one of CUDA's language architects explains what's new, why it's exciting, and demonstrates the power of Kepler GPU accelerators with a real-time cosmology simulation in full 3D.

  Back
 
Keywords:
HPC and AI, Supercomputing 2012 - ID SC2032
Download:
Share:
 
Evolution of GPU Computing
The GPU has evolved rapidly from its invention in 1999 as a VGA Accelerator, becoming a massively parallel general purpose accelerator for heterogeneous computing systems. This talk will focus on significant milestones in GPU architecture and pr ...Read More

The GPU has evolved rapidly from its invention in 1999 as a VGA Accelerator, becoming a massively parallel general purpose accelerator for heterogeneous computing systems. This talk will focus on significant milestones in GPU architecture and programming models, covering several key concepts that demonstrate why advances in GPU-accelerated computing performance and power efficiency will continue to outpace CPUs.

  Back
 
Keywords:
HPC and AI, Supercomputing 2012 - ID SC2033
Download:
Share:
HPC and Supercomputing
Presentation
Media
Titan ORNL's New Computer System for Science
The Oak Ridge Leadership Computing Facility is deploying the Titan supercomputer in support of the U.S. Department of Energys Office of Science programs. This talk will describe the Titan system and its use of NVIDIAs latest GK110 processor. ...Read More

The Oak Ridge Leadership Computing Facility is deploying the Titan supercomputer in support of the U.S. Department of Energys Office of Science programs. This talk will describe the Titan system and its use of NVIDIAs latest GK110 processor.

  Back
 
Keywords:
HPC and Supercomputing, HPC and AI, Supercomputing 2012 - ID SC2005
Download:
Share:
 
Beyond Tsubame 2.0
Tsubame2.0 has been in successful production for the last 2 years, producing numerous research results and accolades. With possible upgrade of the GPUs to Kepler 2s, it will have the capability to surpass the 10 petaflops-class supercomputers in ...Read More

Tsubame2.0 has been in successful production for the last 2 years, producing numerous research results and accolades. With possible upgrade of the GPUs to Kepler 2s, it will have the capability to surpass the 10 petaflops-class supercomputers in single-precision applications, without any increase in the power consumption of 1MW average.

  Back
 
Keywords:
HPC and Supercomputing, Supercomputing 2012 - ID SC2031
Download:
Share:
Medical Imaging and Radiology
Presentation
Media
Fast Simulations of Fast Processes - How GPUs Give Insight Into the Interaction of Light and Matter
Laser-driven radiation sources can potentially help us to cure cancer or understand the dynamics of matter on the atomistic scale. With GPUs we today can simulate these sources at a frames-per-second rate. This in turn enables us to make them af ...Read More

Laser-driven radiation sources can potentially help us to cure cancer or understand the dynamics of matter on the atomistic scale. With GPUs we today can simulate these sources at a frames-per-second rate. This in turn enables us to make them affordable to more users than ever before.

  Back
 
Keywords:
Medical Imaging and Radiology, Supercomputing 2012 - ID SC2023
Download:
Share:
Molecular Dynamics
Presentation
Media
Molecular Dynamics with LAMMPS on a Hybrid Cray Supercomputer
We present software development efforts in LAMMPS that allow for acceleration with GPUs on supercomputers. We present benchmark results for solid-state, biological and mesoscopic systems along with results from simulation of liposomes, polyelect ...Read More

We present software development efforts in LAMMPS that allow for acceleration with GPUs on supercomputers. We present benchmark results for solid-state, biological and mesoscopic systems along with results from simulation of liposomes, polyelectrolyte brushes, and copper nanostructures on graphite. We present methods for efficient simulation with GPUs at larger node counts.

  Back
 
Keywords:
Molecular Dynamics, Supercomputing 2012 - ID SC2024
Download:
Share:
Physics Simulation
Presentation
Media
Quarks, GPUs and Exotic Nuclear Matter
Hadronic matter, such as protons and neutrons, is composed of quarks bound together by gluons whose interactions are described by Quantum Chromodynamics (QCD). In this talk I will describe the use of GPUs for computing the spectrum of QCD. Of sp ...Read More

Hadronic matter, such as protons and neutrons, is composed of quarks bound together by gluons whose interactions are described by Quantum Chromodynamics (QCD). In this talk I will describe the use of GPUs for computing the spectrum of QCD. Of special interest are exotic states, such as those which will be sought by the Glue-X experiment of the Jefferson Lab 12 GeV upgrade, since these states can elucidate the role of the gluons. The calculations range from capacity work on small partitions with a few GPUs to capability sized partitions such as will be available in the Titan system at the Oak Ridge Leadership Facility (OLCF) and I will discuss the work on scaling our application (the Chroma code combined with the QUDA library for QCD on GPUs) to such large systems.

  Back
 
Keywords:
Physics Simulation, Supercomputing 2012 - ID SC2002
Download:
Share:
 
Real-time Triggering Using GPUs in High Energy Physics
In the field of high energy physics, several groups are pursuing the use of GPUs for data analysis and for Monte Carlo simulations of particle interactions. The use of GPUs presented in this seminar is different: GPUs are employed for taking dec ...Read More

In the field of high energy physics, several groups are pursuing the use of GPUs for data analysis and for Monte Carlo simulations of particle interactions. The use of GPUs presented in this seminar is different: GPUs are employed for taking decisions in a trigger system, both as coprocessors in high level software trigger or "embedded" in real-time, fixed-latency hardware trigger.

  Back
 
Keywords:
Physics Simulation, Supercomputing 2012 - ID SC2009
Download:
Share:
 
Many-GPU Calculations in Lattice Quantum Chromodynamics
Lattice Quantum Chromodynamics (QCD) is a computational approach to the theory of the strong nuclear force. State-of-the-art calculations involve integrals over a billion variables or more, which are evaluated using Monte Carlo methods. Such cal ...Read More

Lattice Quantum Chromodynamics (QCD) is a computational approach to the theory of the strong nuclear force. State-of-the-art calculations involve integrals over a billion variables or more, which are evaluated using Monte Carlo methods. Such calculations are typically performed on large-scale distributed systems. In this talk, we outline the main steps in a lattice calculation, and describe multi-GPU implementations of the core routines. We focus in particular on the sparse-matrix linear solves which dominate lattice QCD calculations, and involve significant inter-GPU communication. Preconditioning methods that substantially reduce inter-GPU communication, and hence improve processor utilization, are discussed.

  Back
 
Keywords:
Physics Simulation, Supercomputing 2012 - ID SC2014
Download:
Share:
 
Preparing S3D for Titan from a Domain Scientist's Perspective
Readying S3D, an explicit solver for the compressible reacting Navier-Stokes equations, for Titan took place in conjunction with an effort to move the code from an MPI-everywhere design to a hybrid MPI+x design. The design trade offs and conside ...Read More

Readying S3D, an explicit solver for the compressible reacting Navier-Stokes equations, for Titan took place in conjunction with an effort to move the code from an MPI-everywhere design to a hybrid MPI+x design. The design trade offs and considerations in this process will be described that led to a code ready for large scale GPU computing.

  Back
 
Keywords:
Physics Simulation, HPC and Supercomputing, Supercomputing 2012 - ID SC2028
Download:
Share:
 
Recent Progress on "Path to Exascale" Computational Challenges in Fusion Energy Sciences
Advanced computing is recognized as a vital tool for accelerating progress in scientific research in the 21st Century. The fusion energy research community has made excellent progress in developing advanced codes with associated programming mode ...Read More

Advanced computing is recognized as a vital tool for accelerating progress in scientific research in the 21st Century. The fusion energy research community has made excellent progress in developing advanced codes with associated programming models for which computer runtime and problem size scale well with the number of processors on massively parallel supercomputers. Come see examples of algorithmic progress from the Fusion Energy Sciences area.

  Back
 
Keywords:
Physics Simulation, HPC and Supercomputing, Supercomputing 2012 - ID SC2039
Download:
Share:
Visualization
Presentation
Media
Turbulence Visualization at the Terascale on Desktop PCs
We present a high-performance, throughput-oriented compute system using CUDA to enable the visually guided interactive exploration of large-scale turbulent flows. Our system works on a compressed data representation, employing a wavelet-based co ...Read More

We present a high-performance, throughput-oriented compute system using CUDA to enable the visually guided interactive exploration of large-scale turbulent flows. Our system works on a compressed data representation, employing a wavelet-based compression scheme including run-length and entropy encoding, and efficiently intertwines on-the-fly data decoding and volume ray-casting.

  Back
 
Keywords:
Visualization, Computational Fluid Dynamics, Supercomputing 2012 - ID SC2017
Download:
Share: