The HPC and Supercomputing Track will feature over 90 sessions from industry experts on topics including simulation, visualization, and deep learning. Sessions will focus on how computational and data science are used to solve traditional HPC problems in healthcare, weather, astronomy, and other domains. GPU developers can also connect with innovators and researchers as they share their groundbreaking work using GPU computing.
THE OHIO STATE UNIVERSITY
Professor and Distinguished Scholar /
Research Scientist /
Pushing the Frontier of HPC and Deep Learning
Senior Developer Technology Engineer
Multi-GPU Programming with MPI
UNIVERSITY OF ILLINOIS
Senior Research Programmer
Petascale Molecular Dynamics Simulations from Titan to Summit
NVIDIA, Senior Developer Technology Engineer
Learn how to program multi-GPU systems or GPU clusters using the Message Passing Interface (MPI) and OpenACC or NVIDIA® CUDA®. We'll start with a quick introduction to MPI and how it can be combined with OpenACC or NVIDIA CUDA. Then, we'll cover advanced topics like CUDA-aware MPI and how to overlap communication with computation to hide communication times. We'll also cover the latest improvements with CUDA-aware MPI, interaction with Unified Memory, the multi process service (MPS aka Hyper-Q for MPI), and MPI support in the NVIDIA performance analysis tools.
ABOUT THE SPEAKER: Jiri Kraus is a senior developer in NVIDIA's European DevTech team. In his work, he focuses on multi-GPU programming models and the NVIDIA collaborations with the Juelich Supercomputing Centre.
UNIVERSITY OF ILLINOIS, Senior Research Programmer
The highly parallel molecular dynamics code NAMD is used on the GPU-accelerated Cray XK7 Blue Waters and ORNL Titan machines to perform petascale biomolecular simulations, including a 64-million-atom model of the HIV virus capsid. In 2007, NAMD was one of the first codes to run on a GPU cluster. It’s now being prepared for the ORNL Summit supercomputer, which will feature IBM Power9 CPUs, NVIDIA Volta GPUs, and the NVLink™ CPU-GPU interconnect. Come learn the opportunities and pitfalls of taking GPU computing to the petascale, along with recent NAMD performance advances and early results from the Summit Power8+/P100 "Minsky" development cluster.
ABOUT THE SPEAKER: James Phillips is a Senior Research Programmer in the Theoretical and Computational Biophysics Group at the Beckman Institute for Advanced Science and Technology at the University of Illinois at Urbana-Champaign. He has a Ph.D. in Physics from the University of Illinois. Since 1999, James has been the lead developer of the highly scalable parallel molecular dynamics program NAMD, for which he received a Gordon Bell Award in 2002. His research interests include improving the performance and accuracy of biomolecular simulations through parallelization, optimization, hardware acceleration, better algorithms, and new methods.
Explore new developments in the MVAPICH2-GDR library that help MPI developers realize maximum performance and scalability on HPC clusters with NVIDIA GPUs. See how multiple designs focusing on GPUDirect RDMA(GDR)_Async, non-blocking collectives, support for unified memory, and datatype processing boost HPC application performance. We’ll target emerging deep learning frameworks with novel designs and enhancements to this library to accommodate the large message and dense GPU-computing requirements of the deep learning frameworks. We’ll also present OSU-Caffe—which supports an MPI-based distributed and scalable DL framework—as well as its performance and scalability.
ABOUT THE SPEAKER: Dhabaleswar K. (DK) Panda is a professor and University Distinguished Scholar of Computer Science and Engineering at Ohio State University. He has published over 400 papers in major journals and international conferences. The MVAPICH2 (High-Performance MPI over InfiniBand, iWARP, and RoCE) open-source software package, developed by his research group, is used by more than 2,675 organizations in 83 countries. D.K. is an IEEE fellow and a member of ACM.
Khaled Hamidouche is a research scientist in the Department of Computer Science and Engineering at Ohio State University. His current focus is on designing high-performance unified MPI, PGAS, and hybrid MPI+PGAS runtimes for InfiniBand clusters and their support for accelerators. He has published over 50 papers in international journals and conferences related to these research areas. Khaled has also been actively involved in various professional activities in academic journals and conferences, and is a member of ACM.
Dr. Subramoni is a research scientist in the Department of Computer Science and Engineering at the Ohio State University, USA. His current research interests include high-performance interconnects and protocols, parallel computer architecture, network-based computing, exascale computing, network topology aware computing, QoS, power-aware LAN-WAN communication, fault tolerance, virtualization, big data, and cloud computing.