Thursday, June 24th2:00pm-6:00pmAdvanced MPI Programming details Advanced MPI Programming 2pm - 6pm Pavan Balaji (Argonne National Laboratory), Torsten Hoefler (ETH Zurich), Antonio Peña (Barcelona Supercomputing Center), and Yanfei Guo (Argonne National Laboratory) Abstract zipThe Message Passing Interface (MPI) has been the de facto standard for parallel programming for nearly two decades now. However, a vast majority of applications only rely on basic MPI-1 features without taking advantage of the rich set of functionality the rest of the standard provides. Further, with the advent of MPI-3 (released in September 2012), a vast number of new features have been introduced in MPI, including efficient one-sided communication, support for external tools, non-blocking collective operations, and improved support for topology-aware data movement. The upcoming MPI- 4 standard aims at introducing further improvements to the standard in a number of aspects. This is an advanced-level tutorial that will provide an overview of various powerful features in MPI, especially with MPI-2 and MPI-3, and will present a brief preview into what is being planned for MPI-4. Tutorial Better Scientific Software details Better Scientific Software 2pm - 6pm David E. Bernholdt (Oak Ridge National Laboratory), Anshu Dubey (Argonne National Laboratory), Patricia A. Grubel (Los Alamos National Laboratory), Rinku Gupta (Argonne National Laboratory), and David Rogers (Oak Ridge National Laboratory) Abstract zipThe computational science and engineering (CSE) community is in the midst of an extremely challenging period created by the confluence of disruptive changes in computing architectures, demand for greater scientific reproducibility, and new opportunities for greatly improved simulation capabilities, especially through coupling physics and scales. Computer architecture changes require new software design and implementation strategies, including significant refactoring of existing code. Reproducibility demands require more rigor across the entire software endeavor. Code coupling requires aggregate team interactions including integration of software processes and practices. These challenges demand large investments in scientific software development and improved practices. Focusing on improved developer productivity and software sustainability is both urgent and essential. Tutorial Compression for Scientific & Engineering Data details Compression for Scientific & Engineering Data 2pm - 6pm Franck Cappello (ANL), Peter Lindstrom (Lawrence Livermore National Laboratory), and Sheng Di (Argonne National Laboratory) Abstract Large-scale numerical simulations, observations and experiments are generating very large datasets that are difficult to analyze, store and transfer. Data compression is an attractive and efficient technique to significantly reduce the size of scientific datasets. This tutorial reviews the state of the art in lossy compression of scientific datasets covering the most effective decorrelation, approximation and coding techniques. It details the two leading compressors (SZ and ZFP) that offer lossless and lossy compressions.It also introduces compression error assessment metrics and the Z-checker tool to analyze the difference between initial and decompressed datasets. The tutorial offers hands-on exercises using SZ and ZFP as well as Z-checker. It addresses the following questions: Why compression? How does compression work? How to measure and control compression error of lossy compressors? The tutorial uses examples of real-world compressors and scientific/engineering datasets to illustrate the different compression techniques and their performance. The tutorial is given by two of the leading teams in this domain and targets primarily beginners interested in learning about lossy compression for scientific data. This half-day tutorial is improved from the evaluations of the highly rated tutorials given on this topic at ISC17, SC17, SC18, ISC19, SC19. Tutorial Determining Parallel Application Execution Efficiency and Scaling using the POP Methodology details Determining Parallel Application Execution Efficiency and Scaling using the POP Methodology 2pm - 6pm Judit Giménez (Polytechnic University of Catalonia, Barcelona Supercomputing Center) and Brian J. N. Wylie (Forschungszentrum Jülich GmbH, Jülich Supercomputing Centre) Abstract zipHPC application developers encounter significant challenges getting their codes to run correctly on leadership computer systems consisting of large numbers of interconnected multi-socket multicore processor nodes often with attached accelerator devices. They also need effective tools and methods to track and assess their codes’ execution performance as they aim to get ready for production on current or prospective exascale computer systems. Tutorial Getting Started with Containers on HPC details Getting Started with Containers on HPC 2pm - 6pm Richard S. Canon (Lawrence Berkeley National Lab, NERSC); Andrew J. Younge (Sandia National Labs); Carlos Eduardo Arango Gutierrez (Red Hat, Universidad del Valle); Marco De La Pierre (Pawsey Supercomputing Centre); and Sameer Shende (University of Oregon) Abstract zipWithin just the past few years, the use of containers has revolutionized the way in which industries and enterprises have developed and deployed computational software and distributed systems. The containerization model has gained traction within the HPC community as well with the promise of improved reliability, reproducibility, portability, and levels of customization that were previously not possible on supercomputers. This adoption has been enabled by a number of HPC-compatible Container runtimes that have emerged including Singularity, Shifter, Enroot, Charliecloud, Podman and others. Tutorial Hands-on Practical Hybrid Parallel Application Performance Engineering details Hands-on Practical Hybrid Parallel Application Performance Engineering 2pm - 6pm Christian Feld and Markus Geimer (Jülich Supercomputing Centre); Sameer Shende (University of Oregon; ParaTools, Inc.); and Bill Williams (Technische Universität Dresden, ZIH) Abstract zipThis tutorial presents state-of-the-art performance tools for leading-edge HPC systems founded on the community-developed Score-P instrumentation and measurement infrastructure, demonstrating how they can be used for performance engineering of effective scientific applications based on standard MPI, OpenMP, hybrid combination of both, and increasingly common usage of accelerators. Parallel performance tools from the Virtual Institute - High Productivity Supercomputing (VI-HPS) are introduced and featured in hands-on exercises with Score-P, Scalasca, Vampir, and TAU. We present the complete workflow of performance engineering, including instrumentation, measurement (profiling and tracing, timing and PAPI hardware counters), data storage, analysis, tuning, and visualization. Emphasis is placed on how tools are used in combination for identifying performance problems and investigating optimization alternatives. Using their own notebook computers with a provided E4S [https://e4s.io] OVA image containing all of the necessary tools (running within a virtual machine), participants will conduct exercises on a contemporary HPC system where remote access will be provided for the hands-on sessions through AWS running the E4S image. This will help to prepare participants to locate and diagnose performance bottlenecks in their own parallel programs. Tutorial High Performance Distributed Deep Learning details High Performance Distributed Deep Learning 2pm - 6pm Dhabaleswar Panda (Ohio State University) and Hari Subramoni and Arpan Jain (The Ohio State University) Abstract zipThe recent advances in Deep Learning (DL) have led to many exciting challenges and opportunities for CS and AI researchers alike. Modern DL frameworks like TensorFlow, PyTorch, and several others have emerged that offer ease of use and flexibility to train, and deploy Deep Neural Networks (DNNs). In this tutorial, we will provide an overview of interesting trends in DNN design and how cutting-edge hardware architectures and high-performance interconnects are playing a key role in moving the field forward. We will also present an overview of different DNN architectures and DL frameworks. Most DL frameworks started with a single-node design. However, approaches to parallelize the process of DNN training are also being actively explored. The DL community has moved along different distributed training designs that exploit communication runtimes like MPI and NCCL. We highlight new challenges and opportunities for communication runtimes to exploit high-performance CPU and GPU architectures to efficiently support large-scale distributed DNN training. We also highlight some of our co-design efforts to utilize MPI for large-scale DNN training on cutting-edge CPU and GPU architectures. Finally, we include hands-on exercises to enable the attendees to gain first-hand experience of running distributed DNN training experiments on a modern GPU cluster. Tutorial Kokkos: Performance Portability for C++ Applications and Libraries details Kokkos: Performance Portability for C++ Applications and Libraries 2pm - 6pm Christian R. Trott and Sivasankaran Rajamanickam (Sandia National Labs), Damien Lebrun-Grandie (Oak Ridge National Laboratory), Mikael Simberg (Swiss National Supercomputing Centre), Jonathan Madsen (Lawrence Berkeley National Laboratory), and Geoffrey Womeldorff (Los Alamos National Laboratory) Abstract zipThe trend over the past decade towards increasingly diverse heterogeneous hardware has had a profound impact on the landscape of programming models and execution models. Single-source heterogeneous programming models for performance portability have particularly gained popularity in the past few years due to the increasingly insurmountable plethora of vendor-specific interfaces for heterogeneous hardware. With more than a decade of experience in the field, Kokkos has emerged as a leader in the area of performance-portable parallel programming models, with more than a hundred projects utilizing the latest in CPU and GPU hardware through its generic interface. This tutorial, given in various forms and venues around the globe, presents a practical introduction to the Kokkos programming model. Many iterations on the tutorial material have shown that most attendees acquire the necessary skills to begin porting production applications to the Kokkos programming model after just one day of instruction. Interactive practice exercises teach not only the fundamental syntax and semantics of the programming model, but also the performance characteristics of parallel applications on modern heterogeneous hardware. Presentations of the Kokkos Kernels and Tooling libraries are also included, leaving attendees with all of the necessary skills to use Kokkos in their own HPC applications. Tutorial Managing HPC Software Complexity with Spack details Managing HPC Software Complexity with Spack 2pm - 6pm Todd Gamblin and Gregory Becker (Lawrence Livermore National Laboratory), Massimiliano Culpo (Sylabs), Tamara Dahlgren (Lawrence Livermore National Laboratory), Michael Kuhn (Otto von Guericke University Magdeburg), and Peter Scheibel (Lawrence Livermore National Laboratory) Abstract zipThe modern scientific software stack includes thousands of packages, from C, C++, and Fortran libraries, to packages written in interpreted languages like Python and R. HPC applications may depend on hundreds of packages spanning all of these ecosytems. To achieve high performance, they must also leverage low- level and difficult-to-build libraries such as MPI, BLAS, and LAPACK. Integrating this stack is extremely challenging. The complexity can be an obstacle to deployment at HPC sites and deters developers from building on each others’ work. Spack is an open source tool for HPC package management that simplifies building, installing, customizing, and sharing HPC software stacks. In the past few years, its adoption has grown rapidly: by end-users, by HPC developers, and by the world’s largest HPC centers. Spack provides a powerful and flexible dependency model, a simple Python syntax for writing package build recipes, and a repository of over 3,800 community-maintained packages. This tutorial provides a thorough introduction to Spack’s capabilities: installing and authoring packages, integrating Spack with development workflows, and using Spack for deployment at HPC facilities. Attendees will leave with foundational skills for using Spack to automate day-to-day tasks, along with deeper knowledge for applying Spack to advanced use cases. Tutorial Mastering Tasking with OpenMP details Mastering Tasking with OpenMP 2pm - 6pm Christian Terboven (RWTH Aachen University); Michael Klemm (OpenMP ARB, AMD); Bronis R. de Supinski (Lawrence Livermore National Laboratory); and Xavier Teruel (Barcelona Supercomputing Center) Abstract zipWith the increasing prevalence of multi-core processors, shared-memory programming models are essential. OpenMP is a popular, portable, widely supported and easy-to-use shared-memory model. Since version 3.0 released in 2008, OpenMP offers tasking to support the creation of composable parallel software blocks and the parallelization of irregular algorithms. Developers usually find OpenMP easy to learn. However, mastering the tasking concept of OpenMP requires a change in the way developers reason about the structure of their code and how to expose the parallelism of it. Our tutorial addresses this critical aspect by examining the tasking concept in detail and presenting patterns as solutions to many common problems. Tutorial The Scalable Vector Extension: Programming Tools and Performance Analysis details The Scalable Vector Extension: Programming Tools and Performance Analysis 2pm - 6pm John Linford, Olly Perks, and Roxana Rusitoru (Arm); Simon McIntosh-Smith (University of Bristol); John Levesque (HPE); and Shinji Sumimoto (Fujitsu) Abstract zipThe Scalable Vector Extension (SVE) is the next-generation SIMD instruction set for Armv8-A. SVE does not specify a vector length and it uses predicates to dynamically select vector lanes. For many developers, this presents an entirely new way of thinking about vectorization. This tutorial will introduce tools from the Arm community for SVE programming and performance analysis, i.e. compilers, scientific libraries, profilers, and debuggers. The content of this tutorial will be notably different from past SVE tutorials as this will be the first time a public SVE tutorial will be taught on SVE hardware. We plan to provide remote access to SVE-enabled CPUs for this tutorial. Hands-on exercises will explore the unique features of SVE and demonstrate their applicability to a range of common programing motifs. This tutorial will demonstrate how to capture high-utility information and present it in meaningful ways, i.e. how much time is spent in application routines, where these routines are called in the source code, and how well the routines vectorize. Attendees will complete the tutorial with a working understanding of SVE and knowledge of how SVE may be used in their applications. Tutorial | Friday, June 25th2:00pm-6:00pmHands-On HPC Application Development Using C++ and SYCL details Hands-On HPC Application Development Using C++ and SYCL 2pm - 6pm Gordon Brown and Michael Wong (Codeplay Software), Michael Steyer (Intel), Rod Burns (Codeplay Software), Aksel Alpay (Heidelberg University), Peter Zuzek (Codeplay Software), Ronan Keryell (Xilinx), and Igor Vorobtsov (Intel) Abstract zipSYCL is a programming model that lets developers write parallel applications for a wide variety of devices (CPUs, GPUs, FPGAs and more) from a single code base using standard C++. Given the growing heterogeneity of processor roadmaps, moving to a platform-independent model such as SYCL is essential for modern software developers. SYCL has the further advantage of supporting a single-source style of programming from completely standard C++. In this tutorial, we will introduce SYCL and provide programmers with a solid foundation they can build on to gain mastery of this language. The real value in SYCL comes from its use in writing applications. We will explore how SYCL can be used to write serious applications, covering intermediate to advanced features of SYCL as well as some of the tools and libraries that support SYCL application development. This is a hands-on tutorial. Students will be given exercises that represent key design patterns encountered by people who program heterogeneous systems. Tutorial InfiniBand, High-speed Ethernet, RoCE, Omni-Path, EFA, and Slingshot for Beginners details InfiniBand, High-speed Ethernet, RoCE, Omni-Path, EFA, and Slingshot for Beginners 2pm - 6pm Dhabaleswar Panda and Hari Subramoni (The Ohio State University) Abstract zipInfiniBand (IB), High-speed Ethernet (HSE), RoCE, Omni-Path, EFA, and Slingshot technologies are generating a lot of excitement towards building next generation High-End Computing (HEC) systems including clusters, datacenters, file systems, storage, cloud computing and Big Data (Hadoop, Spark, HBase and Memcached) environments. This tutorial will provide an overview of these emerging technologies, their offered architectural features, their current market standing, and their suitability for designing HEC systems. It will start with a brief overview of IB, HSE, RoCE, Omni-Path, EFA, Slingshot, and Omni-Path. In-depth overview of the architectural features of IB, HSE (including iWARP and RoCE), and Omni-Path, their similarities and differences, and the associated protocols will be presented. An overview of the emerging NVLink, NVLink2, and NVSwitch architectures will also be given. Next, an overview of the OpenFabrics stack which encapsulates IB, HSE, and RoCE (v1/v2) in a unified manner will be presented. An overview of libfabrics stack will also be provided. Hardware/software solutions and the market trends behind these networking technologies will be highlighted. Sample performance numbers of these technologies and protocols for different environments will be presented. Finally, hands-on exercises will be carried out for the attendees to gain first-hand experience of running experiments with high-performance networks. Tutorial Introduction to HPC: Applications, Systems, and Programming Models details Introduction to HPC: Applications, Systems, and Programming Models 2pm - 6pm Bernd Mohr (Jülich Supercomputing Centre (JSC)) Abstract In this introductory tutorial, you will learn what "high performance computing" means and what differentiates it from more mainstream areas of computing. You will also be introduced to the major applications that use high performance computing for research and commercial purposes, and how AI and HPC interact with each other. Then, we present the major HPC system architectures needed to run these applications. Finally, you will be provided with an overview of the languages and paradigms used to program HPC applications and systems. The tutorial will be presented by Dr.-Ing. Bernd Mohr from the Jülich Supercomputing Centre. Tutorial Maintaining a Modern Scientific Software Stack Made Easy with EasyBuild details Maintaining a Modern Scientific Software Stack Made Easy with EasyBuild 2pm - 6pm Kenneth Hoste (Ghent University), Alan O'Cais and Markus Geimer (Forschungszentrum Juelich GmbH), and Bart Oldeman (Compute Canada) Abstract zipInstalling scientific software for supercomputers is known to be a tedious and time-consuming task. Increasing HPC user community diversity, as well as hardware complexity, continues to deepen the application software stack. Delivering optimised software installations and providing access to these installations in a reliable, user-friendly way is a highly non-trivial task that affects the application developer, the HPC user support teams, and the users themselves. This tutorial addresses this issue, providing the attendees with the knowledge to install and manage software in a clean and reproducible way. For this, the tutorial introduces EasyBuild, a software build and installation framework implemented in Python. It supports implementing (complex) installation procedures concisely, and is able to fully autonomously perform optimised software installations. It is open source and has a thriving community, and supports more than 2,200 (scientific) software packages. It is well integrated with different implementations of environment modules (Lmod, Tcl, C) to generate module files automatically and present them in a consistent way. Tutorial Modern Mixed- and Multi-Precision Methods details Modern Mixed- and Multi-Precision Methods 2pm - 6pm Hartwig Anzt (Karlsruhe Institute of Technology, University of Tennessee); Jack Dongarra (University of Tennessee, Oak Ridge National Laboratory); and Piotr Luszczek (University of Tennessee, Tickle College of Engineering) Abstract zipThis tutorial will expose the audience to the rapidly expanding landscape of mixed- and multi-precision methods. The ongoing cross-pollination between high-performance computing (HPC) and machine learning (ML) is leading to intelligent computational steering of large-scale simulations. More importantly for this tutorial, sharing of the hardware platforms and exploiting their wide range of computational modes has led to proliferation of multiple representations of floating-point data—and increased interest in new methods that exploit them. Against the backdrop of high-performance libraries produced by internet-scale companies, hardware vendors, national laboratories, and academic institutions, we will show the recent algorithmic progress in exploiting multiple precisions for increased efficiency in performance, communication, and/or storage. The techniques presented in the tutorial employ floating-point representations such as limited precision, quantized integers, and modular precision ecosystems. A portion of this tutorial will cover the fundamentals of the HPC software development in order to introduce the audience to some of the low-level details of coding for multiple precisions on modern hardware, including accelerators. The hardware focus of the tutorial will feature floating-point representation and performance of the recent supercomputing and industrial computing CPUs and accelerators, as well as less mainstream devices meant for more specific tasks at limited power envelopes. Tutorial OpenMP Common Core: Learning Parallelization of Real Applications from the Ground-Up details OpenMP Common Core: Learning Parallelization of Real Applications from the Ground-Up 2pm - 6pm Manuel Arenaz (Appentra Solutions, University of A Coruña); Barbara Chapman (Brookhaven National Lab); Oscar Hernandez (NVIDIA); Reuben D. Budiardja (Oak Ridge National Lab (ORNL)); and Dossay Oryspayev (Brookhaven National Lab) Abstract zipAs HPC continues to move towards a model of multicore and accelerator programming, a detailed understanding of shared-memory models and how best to use accelerators has never been more important. OpenMP is the de facto standard for writing multithreaded code to take advantage of shared memory platforms, but to make optimal use of it can be incredibly complex. Tutorial Productive Parallel Programming for FPGA with High-Level Synthesis details Productive Parallel Programming for FPGA with High-Level Synthesis 2pm - 6pm Johannes de Fine Licht and Torsten Hoefler (ETH Zurich) Abstract zipEnergy efficiency has become a first class citizen in the design of large computing systems. While GPUs and custom processors show merit in this regard, reconfigurable architectures, such as FPGAs, promise another major improvement in energy efficiency, constituting a middle ground between fixed hardware architectures and custom-built ASICs. This tutorial shows how high-level synthesis (HLS) can be harnessed to productively achieve scalable pipeline parallelism on FPGAs. Attendees will learn how to target FPGA resources from high-level C++ or OpenCL code, guiding the mapping from imperative code to hardware, enabling them to develop massively parallel designs. We treat well-known examples from the software world, relating traditional code optimizations to hardware, building on existing knowledge when introducing new topics. By bridging the gap between software and hardware optimization, our tutorial aims to enable developers from a large set of backgrounds to start tapping into the potential of FPGAs for high performance codes. Tutorial | Friday, July 2nd2:00pm-6:00pm16th Workshop on Virtualization in High-Performance Cloud Computing details 16th Workshop on Virtualization in High-Performance Cloud Computing 2pm - 6pm Michael Alexander (BOKU, Vienna); Anastassios Nanos (Nubificus Ltd.); Anastasios Nanos and Charalampos Mainas (Nubificus LTD); Panagiotis Kokkinos (ICCS/NTUA); Yiannis Gkoufas and David Yu Yuan (IBM Research, Ireland); Sezar Jarrous-Holtrup (University of Münster,); Folker Schamel (Spinor GmbH, Germany); Argyrios Kokkinis (Aristotle University of Thessaloniki, Greece); Achilleas Tzenetopoulos (National Technical University of Athens); Remo Andreoli (Scuola Superiore Sant'Anna, Italy); Ricardo Rocha and Maria Girone (CERN); and Stefan Hajnoczi (Red Hat Research) Abstract pdfWorkshop Website https://vhpc.orgContainers and virtualization technologies constitute key enabling factors for flexible resource management in modern data centers, and particularly in cloud environments. HPC operators and cloud providers need to manage complex infrastructures in a seamless fashion to support the highly dynamic and heterogeneous workloads and hosted applications customers deploy. Various virtualization-containerization technologies contribute to the overall picture in different ways: machine virtualization, with its capability to enable consolidation of multiple underutilized servers with heterogeneous software and operating systems (OSes), and its capability to live-migrate a fully operating virtual machine (VM) with a very short downtime, enables novel and dynamic ways to manage physical servers; OS-level virtualization (i.e., containerization), with its capability to isolate multiple user-space environments and to allow for their coexistence within the same OS kernel, promises to provide many of the advantages of machine virtualization with high levels of responsiveness and performance; lastly, unikernels provide for many virtualization benefits with a minimized OS/library surface. The Workshop on Virtualization in High-Performance Cloud Computing (VHPC) aims to bring together researchers and industrial practitioners facing the challenges posed by virtualization, containerization, and orchestration techniques for deploying HPC and HPC-like applications in large data centers. Workshop 2nd ISC-HPC International Workshop on “Monitoring and Operational Data Analytics” (MODA) details 2nd ISC-HPC International Workshop on “Monitoring and Operational Data Analytics” (MODA) 2pm - 6pm Florina Ciorba (University of Basel); Daniele Tafani (Fujitsu, Germany); Utz-Uwe Haus (Cray EMEA Research Lab); Nicolas Lachiche (University of Strasbourg); Ann Gentile (Sandia National Laboratories); Natalie Bates (Energy Efficiency HPC WG); Torsten Wilde (HPE); Aleš Zamuda (University of Maribor); Martin Molan (University of Bologna); and Masaaki Terai (RIKEN Centre for Computational Science) Abstract pdfWorkshop Website https://moda21.sciencesconf.org/This workshop aims to provide insight into the current state and trends in Monitoring and Operational Data Analytics (MODA), to identify potential gaps, and to offer an outlook into the future of MODA at a large scale together with possible solutions for the upcoming Exascale systems. The focus of the MODA21 workshop will be on currently envisioned solutions and best practices for monitoring systems at data centers and HPC sites, as well as on effective strategies for analyzing and interpreting the collected operational data. The workshop is unique to the European HPC arena in addressing these topics. We envision a balanced mix between a keynote address, peer-reviewed technical paper presentations, invited talks, and a panel discussion. Workshop 2nd International Workshop on Machine Learning Hardware details 2nd International Workshop on Machine Learning Hardware 2pm - 6pm Pete Beckman and Swann Perarnau (Argonne National Laboratory), Rosa M. Badia (Barcelona Supercomputing Center), Kentaro Sano (RIKEN), Valentin Reis (Groq), Prasanna Balaprakash (ANL), Rob Schreiber (Cerebras), Tony Hey (STFC), Michaela Blott (Xilinx), Tianjian Lu (Google Research), Haohuan Fu (Tsinghua University), and Takano Ryousei (AIST) Abstract pdfWorkshop Website https://mlhardware.github.ioRecent years have seen a surge of investment in AI chip companies worldwide. These companies are however mostly targeting applications outside of the scientific computing community. As the use of ML accelerates in the HPC field itself, there is concern that the scientific community should influence the design of this new specialized hardware. Indeed, scientific computing has a distinctive set of requirements regarding workload type, usage model, and platform administration. How those chips answer those demands will shape the future of their integration within the global scientific computing infrastructure. In this workshop, we propose to let the community and select vendors engage on questions related to the low-level aspects of this new hardware and its integration with HPC systems, as well as to the software APIs and compiler toolchains that will be available. This proposal follows the outcome of a successful ISC20 workshop where an emphasis on compiler technology emerged. In this iteration, we will push the agenda further by discussing existing efforts to leverage ML hardware for scientific HPC applications. Workshop 5th International Workshop on In Situ Visualization details 5th International Workshop on In Situ Visualization 2pm - 6pm Tom Vierjahn (Westphalian University of Applied Sciences), Thomas Theussl (KAUST Core Labs), Steffen Frey (University of Groningen), Kenneth Moreland (Sandia National Laboratories), Guido Reina (University of Stuttgart), and Andrew Bauer (United States Army Corps of Engineers) Abstract pdfWorkshop Website https://woiv.gitlab.ioThe ever increasing scale of today’s HPC simulations with their inherent I/O bottleneck makes in situ an essential approach for data analysis. Nowadays, the rate of data generation easily exceeds available bandwidth or storage capabilities significantly. Consequently, analysis and visualization has to be coubled in situ to a live simulation in order to facilitate comprehensive investigation. In doing so, data abstracts are generated that capture much more information than otherwise possible. Workshop 7th Annual High Performance Container Workshop details 7th Annual High Performance Container Workshop 2pm - 6pm Christian Kniep (AWS); Shane Canon (Lawrence Berkeley National Labs); Andrew Younge (Sandia National Laboratories); Carlos Eduardo Arango Gutierrez (Red Hat); Abdulrahman Azab (University of Oslo, Partnership for Advanced Computing in Europe (PRACE)); Umesh Upadhyaya (HPC Nepal); Michael Kuhn (Otto von Guericke University Magdeburg); and Carsten Kutzner (Max Planck Institute for Biophysical Chemistry) Abstract pdfWorkshop Website https://hpcw.github.ioLinux Containers have become an industry standard for sharing and distributing software. This workshop will create a space of interaction between field experts,end-users, and newcomers to discuss container technologies and their current and future challenges. Workshop Approximate and Transprecision Computing on Emerging Technologies (ATCET) - Second Edition details Approximate and Transprecision Computing on Emerging Technologies (ATCET) - Second Edition 2pm - 6pm A. Cristiano I. Malossi (IBM Research - Zurich); Luca Benini (Dep. of Inform.Technol. Electrical Eng., ETH Zurich, Switzerland); Norbert Wehn (Dep. of Electrical and Computer Engineering, Technische Universität Kaiserslautern, Germany); Roger Woods (School of Electronics, Electrical Engineering and Computer Science, Queen's University, Belfast, United Kingdom); Andrew Emerson (Department of High Performance Computing, CINECA, Italy); Frank K. Gurkaynak (ETH Zürich); Alberto Bosio (École Centrale de Lyon); Dimitrios S. Nikolopoulos (Virginia Tech); Christos-Savvas Bouganis (Imperial College); and Enrique Salvador Quintana Ortí (Universitat Politècnica de València) Abstract pdfWorkshop Website http://oprecomp.eu/atcet-2021Guaranteed numerical precision of each elementary step in a complex computation has been the mainstay of traditional computing systems for many years. This era is at its twilight: to overcome the “power wall” in Exascale systems, a shift from traditional computing paradigms is now mandatory. This workshop will investigate the theoretical and practical understanding of the energy efficiency boost obtainable when accuracy requirements on data being processed, stored and communicated can be lifted for intermediate calculations. The target applications range from Big Data Analytic and Deep Learning, up to classical scientific computing simulations in HPC environments. Workshop Arm HPC User Group (AHUG) 2021 details Arm HPC User Group (AHUG) 2021 2pm - 6pm Jonathan C. Beard (Arm Research), Jeffrey Young (Georgia Tech), Roxana Rusitoru (Arm Research), Oscar Hernandez (NVIDIA), Andrew Younge (Sandia National Lab), RuQing (G.) Xu Xu (University of Tokyo), Yuetsu Kodama (RIKEN R-CCS), Steve Messenger (Amazon), Srinath Vadlamani (Arm), Miwako Tsuji (RIKEN R-CCS), Stepan Nassyr (Juelich Supercomputing Centre), Federica Filippini (Politecnico di Milano), Andrei Poenaru (University of Bristol), John Linford (Arm), Jeff Hammond (NVIDIA), Luigi Genovese (Atomistic Simulation Laboratory (L_Sim) - CEA Grenoble), Miquel Moreto (Barcelona Supercomputing Center), Hatem Ltaief (KAUST), Bine Brank (Juelich Supercomputing Centre), Eva Siegmann (Stony Brook University), Sarat Sreepathi (Oak Ridge National Laboratory), Andrew Younge (Sandia National Laboratories), Thomas Chen (U.S. Technology Policy Committee), and Craig Prunty (SiPearl) Abstract pdfWorkshop Website https://a-hug.org/isc-2021-event/After the success and great interest from the last 3 years, Arm HPC User's Group at ISC will be bringing an even more diverse and exciting panel of topics ranging from the latest Arm-based systems, to programming for arm, co-design, to new HPC areas, such as deep learning, edge, and data-center analytics. Workshop Compiler-assisted Correctness Checking and Performance Optimization for HPC details Compiler-assisted Correctness Checking and Performance Optimization for HPC 2pm - 6pm Emmanuelle Saillard (Inria), Julien Jaeger (CEA), Ira Baxter (Semantic Designs), Tim Jammer (TU Darmstadt), Julia Lawall (Inria), Michael Blesel (Otto-von-Guericke-University Magdeburg), and Reed Milewicz (Sandia National Laboratories) Abstract pdfWorkshop Website https://c3po-workshop.github.io/2021/indexPractical compiler-enabled programming environments, applied analysis methodologies, and end-to-end toolchains can contribute significantly to performance portability in the exascale era. The practical and applied use of compilation techniques, methods, and technologies, including static analysis and transformation, are imperative to improve the performance, correctness, and scalability of high-performance applications, middleware, and reusable libraries. This workshop brings together a diverse group of researchers with a shared interest in applying compilation and source-to-source translation methodologies, among others, to enhance explicit parallel programming such as MPI, OpenMP, and hybrid models. These types of compiler technologies can also be applied to heterogeneous programming elements including FPGAs and GPUs in order to deliver higher achievable performance as compared to library-based methods and human-coded approaches taken in isolation. Workshop Deep Learning on Supercomputers details Deep Learning on Supercomputers 2pm - 6pm Valeriu Codreanu (SURFsara); Ian T. Foster (University of Chicago, Argonne National Laboratory); Zhao Zhang (Texas Advanced Computing Center); Torsten Hoefler (ETH Zurich); Arvind Ramanathan (Argonne National Laboratory); Alexandre Bonvin and Manon Réau (Utrecht University); Stefan Kesselheim (Jülich Supercomputing Center); Jonas Teuwen (Netherlands Cancer Institute); and Chen Liu (SambaNova Systems) Abstract pdfWorkshop Website https://dlonsc.github.io/The Deep Learning (DL) on Supercomputers workshop provides a forum for practitioners working on any and all aspects of DL for science and engineering in the High Performance Computing (HPC) context to present their latest research results and development, deployment, and application experiences. The general theme of this workshop series is the intersection of DL and HPC; the theme of this particular workshop is the applications of DL methods in science and engineering: novel uses of DL methods, e.g., convolutional neural networks (CNN), recurrent neural networks (RNN), generative adversarial networks (GAN), and reinforcement learning (RL), in the natural sciences, social sciences, and engineering, to enhance innovative applications of DL in traditional numerical computation. Its scope encompasses application development in scientific scenarios using HPC platforms; DL methods applied to numerical simulation; fundamental algorithms, enhanced procedures, and software development methods to enable scalable training and inference; hardware changes with impact on future supercomputer design; and machine deployment, performance evaluation, and reproducibility practices for DL applications with an emphasis on scientific usage. This workshop will be centered around published papers. Submissions will be peer-reviewed, and accepted papers will be published as part of the Joint Workshop Proceedings by Springer. Workshop Fourth HPC Applications in Precision Medicine Workshop details Fourth HPC Applications in Precision Medicine Workshop 2pm - 6pm Eric Stahlberg (Frederick National Laboratory for Cancer Research), Thomas Steinke (Zuse Institute Berlin), Jan Nygard (The Cancer Registry of Norway), Marco Berghoff (Karlsruhe Institute of Technology), Sunita Chandrasekaran (University of Delaware), Petrina Hollingsworth (Frederick National Laboratory for Cancer Research), and Andreas Deutsch (Dresden University of Technology) Abstract pdfWorkshop Website https://ncihub.org/groups/hapm21High-performance computing has become central to the future success of precision medicine. Catalyzed by the dramatic increase in the volume of research and clinical data available through sequencing and advanced imaging techniques, clinical data available through medical records and mobile health devices, research data from automated platforms, combined with molecular and multi-scale simulations and the rapid adoption of deep learning approaches has created a convergence shaping the frontiers of computing and precision medicine. New approaches to drug discovery, data sharing, aggregation and safeguards, use of machine learning models in research and clinical contexts, multi-scale observations integrated with predictive modeling, and biomedical digital twins have identified new challenges and opportunities in these rapidly evolving frontiers. Workshop HPC I/O in the Data Center details HPC I/O in the Data Center 2pm - 6pm Julian Kunkel (University of Reading), Jay Lofstead (Sandia National Laboratories), Jean-Thomas Acquaviva (Data Direct Networks), Bingsheng He (National University of Singapore), Richard Lawrence (MetOffice), Erdem Yilmaz (University of Reading), Glenn Lockwood (Lawrence Berkeley National Laboratory), Frank Gadban (University of Hamburg), Luke Logan (Illinois Tech), Jack Kolokasis (Foundation for research and Technology Hellas), Alberto Scionti (Linksfoundation), and Bruno Silva (Amazon) Abstract pdfWorkshop Website https://hps.vi4io.org/events/2021/iodcManaging scientific data at scale is challenging for scientists but also for the host data center. The storage and file systems deployed within a data center are expected to meet users' requirements for data integrity and high performance across heterogeneous and concurrently running applications. Workshop ISC'21 SuperCompCloud: 4th International Workshop on Interoperability of Supercomputing and Cloud... details ISC'21 SuperCompCloud: 4th International Workshop on Interoperability of Supercomputing and Cloud Technologies 2pm - 6pm Sadaf Alam (Swiss National Supercomputing Centre); David Hancock (Indiana University); François Tessier (INRIA); David Y. Hancock (Indiana University); Nic Bellingham (UK Met Office); Stig Telfer (StackHPC); Andrew Jones (MS Azure); Mallikarjun (Arjun) Shankar (ORNL); Bjoern Enders and Gabor Torok (NERSC, LBNL); and Christopher Haine (HPE HPC/AI EMEA Research Lab) Abstract pdfWorkshop Website https://sites.google.com/view/supercompcloud/isc21-4th-supercompcloud-workshopImminent arrival and deployment of exascale systems among multiple hyperscaler cloud providers are expected to enable breakthroughs for various scientific disciplines. Increasingly, these systems utilize cloud technologies, enabling complex and distributed workflows that improve not only scientific productivity, but accessibility of resources to a wide range of communities. Such an integrated and seamlessly orchestrated system for supercomputing and cloud technologies is indispensable for experimental facilities that have been experiencing an unprecedented rate of data growth. While limited-scale HPC services has been available in public cloud environments, petascale and beyond data and computing capabilities are provisioned within HPC data centers using close-to-metal provisioning services to ensure performance, scaling, and cost effectiveness. This workshop aims to bring together experts and practitioners from academia, national laboratories and industry to discuss technologies, use cases and best practices in order to set a vision and direction for leveraging extreme-scale computing and on-demand cloud ecosystems. Workshop Machine Learning on HPC Systems details Machine Learning on HPC Systems 2pm - 6pm Janis Keuper (Fraunhofer Institut für Techno- und Wirtschaftsmathematik ITWM; Institute for machine Learning and Analytics (IMLA), Offenburg University); Sunna Torge (TU Dresdden); Jenia Jitsev (Institute for Advanced Simulation (IAS)); Juan Jose Durillo (Leibniz Supercomputing Centre); Dennis Hoppe (HLRS); Daniel Soudry (Technion); Jenia Jitsev (Juelich Supercomputer Center); Nico Hoffmann (TU Dresden); Stefanie Günther (LLNL); Ahmed Elnaggar (TU Munich); Kalun Ho (Fraunhofer Center HPC); Niranjan Hasabnis (Intel); Mathias Esteban and Jonathan Muraña (Universidad de la República); and Peter Winkler (TU Dresden) Abstract pdfWorkshop Website http://www.MLHPCS.orgOver the last few years, Machine Learning (and in particular Deep Learning) (ML / DL) has become an important research topic in the High Performance Computing (HPC) community. This comes along with new users and data intensive applications on HPC systems, which increasingly affects the design and operation of compute infrastructures. Bringing new users and data intensive applications on HPC systems, Learning methods are increasingly affecting the design and operation of compute infrastructures. On the other hand, the learning community is just getting started to utilize the performance of HPC, leaving many opportunities for better parallelization and scalability. The intent of this workshop is to bring together researchers and practitioners from all communities to discuss three key topics in the context of High Performance Computing and learning methods: parallelization and scaling of ML / DL algorithms, learnig applications on HPC systems, and HPC systems design and optimization for ML / DL workloads. Workshop Numerical Algorithms and Libraries for Exascale details Numerical Algorithms and Libraries for Exascale 2pm - 6pm Hatem Ltaief, Bilel Hadri, and David Keyes (KAUST); Daniel Grünewald (Fraunhofer ITWM); Laura Grigori (INRIA); Alfredo Buttari (CNRS-IRIT); Hartwig Anzt (Karlsruhe Institute of Technology); Ulrike Yang (Lawrence Livermore National Laboratory); and John Shalf (Lawrence Berkeley National Laboratory) Abstract pdfWorkshop Website https://cemse.kaust.edu.sa/hicma/events/event/isc21-workshop-numerical-algorithms-and-libraries-exascale-nal-xWith the hardware technology scaling and the trend on heterogeneous chip design, the existing numerical algorithms and software framework may break down due to load imbalance. There is currently a fundamental mismatch between the underlying hardware architecture with high thread concurrency and the software deployment of numerical libraries, which relies on the traditional bulk synchronous programming model. Numerical software should first squeeze performance out of single node by efficiently running on manycore architectures with processor counts sharing a common memory in the hundreds. Programming and extracting performance from these advanced architecture chips remain a challenging effort, which is further exacerbated in distributed-memory environment. Algorithmic solutions such as fine-grained computations, communication/synchronization-reducing, and mixed precisions come to the rescue. They represent some of the key ingredients to embrace for software libraries moving forward and leverage extreme-scale computing capabilities. Workshop Second International Workshop on the Application of Machine Learning Techniques to Computational... details Second International Workshop on the Application of Machine Learning Techniques to Computational Fluid Dynamics and Solid Mechanics Simulations and Analysis 2pm - 6pm Volodymyr Kindratenko (National Center for Supercomputing Applications, University of Illinois at Urbana-Champaign); Andreas Lintermann (Jülich Supercomputing Centre, Forschungszentrum Jülich); Charalambos Chrysostomou (The Cyprus Institute); Ashley Scillitoe (The Alan Turing Institute); Eloisa Bentivegna (IBM Research Europe); Jiahuan Cui (Zhejiang University); George Karniadakis (Brown University, MIT); Kazuto Ando (RIKEN Center for Computational Science (R-CCS)); Mario Rüttgers (Institute of Aerodynamics and Chair of Fluid Mechanics, RWTH Aachen University; Jülich Supercomputing Centre, FZ Jülich); Michael Gauding (CORIA and University of Rouen); Mario Bedrunka and Felipe de Castro Teixeira Carvalho (University of Siegen); Jinhui Yan (University of Illinois at Urbana-Champaign); Thomas Brown (George Mason University, Center for Mathematics and Artificial Intelligence, Center for Computational Fluid Dynamics); Sarath Radhakrishnan (Barcelona Supercomputing Center); Yaning Wang (Zhejiang University); and Shirui Luo (University of Illinois at Urbana-Champaign) Abstract pdfWorkshop Website http://www.ncsa.illinois.edu/enabling/data/deep_learning/news/cfdml21Combination of computational fluid dynamics (CFD) with machine learning (ML) is a newly emerging research direction with the potential to enable solving so far unsolved problems in many application domains. This workshop aims to demonstrate the use of high-fidelity CFD simulations to generate data and utilize it to train ML models to better predict the underlying physics in fluid dynamics utilizing the breakthrough in computational power, the evolution of data science techniques, and the ability to generate terabytes of data from high-fidelity simulations. ML techniques have the potential to support the identification and extraction of hidden features in large-scale flow computations, hence allowing to shift the focus from time-consuming feature detection to in-depth examinations of such features. Furthermore, ML techniques have the ability to find undetected correlations between phenomena in the flow, which will lead to deeper insight of the physics involved in complex natural processes. Apart from pure fluid dynamic, other research areas, such as constitutive modeling of heterogeneous materials, multiphase flow modelling, dynamics of the atmospheric, ocean, and climate system, and combustion/chemical reactions are working on similar techniques. The workshop will stimulate this research by providing a venue to exchange new ideas and discuss challenges and opportunities. Workshop Sixth International Workshop on Communication Architectures for HPC, Big Data, Deep Learning and... details Sixth International Workshop on Communication Architectures for HPC, Big Data, Deep Learning and Clouds at Extreme Scale 2pm - 6pm Dhabaleswar Panda, Hari Subramoni, and Aamir Shafi (The Ohio State University); Satoshi Matsuoka (RIKEN Center for Computational Science); Gilad Shainer (NVIDIA/Mellanox); Duncan Roweth (HPE); Phil Murphy (Cornelius Networks); Hemal Shah and Moshe Voloshin (Broadcom); Matthew Williams (Rockport Networks); and Sanjay Basu (Oracle) Abstract pdfWorkshop Website http://nowlab.cse.ohio-state.edu/exacomm/Extreme-Scale Computing in HPC, Big Data, Deep Learning, and Clouds are marked by multiple-levels of hierarchy and heterogeneity ranging from the compute units (many-core CPUs, GPUs, APUs, etc) to storage devices (NVMe, NVMe over Fabrics etc) to the network interconnects (InfiniBand, High-Speed Ethernet, Slingshot, etc). Owing to the plethora of heterogeneous communication paths with different cost models expected to be present in extreme-scale systems, data movement is seen as the soul of different challenges for exascale computing. On the other hand, advances in networking technologies such as NoCs (like NVLink and Stormlake), emergence of new I/O interface architecture standards (CCIX, Gen-Z, CAPI etc), and RDMA enabled networks and the likes are constantly pushing the envelope of research in the field of novel communication and computing architectures for extreme-scale computing. The goal of this workshop is to bring together researchers and software/hardware designers from academia, industry and national laboratories who are involved in creating network-based computing solutions for extreme-scale architectures. The scope of the workshop includes, but not limited to: scalable communication architectures and protocols, high performance networks, runtime/middleware designs, novel hardware/software co-design, high performance communication solutions for accelerator-based computing, power-aware techniques and designs, performance evaluations, QoS, and virtualization. Workshop The Second Workshop on LLVM Compiler and Tools for HPC details The Second Workshop on LLVM Compiler and Tools for HPC 2pm - 6pm Johannes Doerfert (Argonne National Laboratory), Anja Gerbes (Center for Information Services and High Performance Computing), Sameer Shende (University of Oregon), Jeremy Bennett (Embecosm), Shintaro Iwasaki (Argonne National Laboratory), Ernesto Su and Xinmin Tian (Intel), Valentin Clement (Oak Ridge National Laboratory), Arnamoy Bhattacharyya (Huawei), Saiyedul Islam (Advanced Micro Devices (AMD)), Tobias Grosser (University of Edinburgh), and Jeffrey Sandoval (HPE) Abstract pdfWorkshop Website https://hps.vi4io.org/events/2021/llvmThe LLVM framework is a vast ecosystem surrounding a compiler core which enabled various advances in source-code tools, debuggers, linkers, and a whole host of programming-language and toolchain-related components. In addition to the open source components in the LLVM framework, e.g., Clang, Flang, MLIR, LLDB, etc., LLVM also serves as a foundation for a majority of vendor compilers and toolchains. As such, most current and future HPC system will come with LLVM components that are developed in large parts in the open source LLVM code base, a multi-company, multi-research institute, industry-academia, joint-venture. Workshop Third Workshop on HPC Education and Training for Emerging Technologies details Third Workshop on HPC Education and Training for Emerging Technologies 2pm - 6pm Nitin Sukhija (Slippery Rock University Of Pennsylvania); Scott Lathrop (University of Illinois); Nia Alexandrov (Hartree Centre, STFC, UKRI); Andrew Jones (Microsoft); Marjut Dieringer and Kevin McFall (NVIDIA); Kjartan Thor Wikfeldt (EuroCC National Competence Center Sweden); Mozhgan Chimeh (NVIDIA); Weronika Filinger (Edinburgh Parallel Computing Centre (EPCC)); Julia Mullen (MIT Lincoln Laboratory); Ann Backhaus (Pawsey Supercomputing Centre); and Richard Lawrence (Texas A&M University) Abstract pdfWorkshop Website https://sighpceducation.acm.org/events/HETET21.htmlHPC is central for empowering progress in diverse scientific and non-scientific domains. A myriad of technologies in the post peta-scale computing demand a significantly greater degree of parallelism than we currently observe. The rapid advancement of new HPC technologies has facilitated the convergence of Artificial Intelligence (AI), Big Data Analytics, and the HPC platforms to solve complex, large-scale, real-time analytics and applications for scientific and non-scientific fields. As we move towards exascale, the convergent computing platforms along with a paradigm shift in the programming applications for them provide both challenges and opportunities, for cyberinfrastructure facilitators and educators to prepare and support a diverse community of professionals to utilize evolving HPC, equipping them to solve complex scientific, engineering, and technological problems. Workshop Workshop on the In Situ Co-Execution of High-Performance Computing & Data Analysis details Workshop on the In Situ Co-Execution of High-Performance Computing & Data Analysis 2pm - 6pm Julien Bigot (Commissariat à l'énergie atomique et aux énergies alternatives), Bruno Raffin (Inria), Leonardo Bautista Gomez (Barcelona Supercomputing Center), Wounter Klinj (Forschungszentrum Julich), Charles Gueunet (Kitware), Sai Narasimhamurthy (Seagate Systems), Achim Basermann (DLR), Matthieu Dorier (ANL), Amal Gueroudji (CEA), Tiago Quintino (ECMWF), Dirk Pleiter (Forschungszentrum Julich), Alejandro Ribes (EDF), Virginie Grandgirard (CEA), and Yuuichi Asahi (JAEA) Abstract pdfWorkshop Website https://hpcda.github.io/Exascale promises to support disruptive numerical experiments generating unprecedented quantity and quality of data. Only the latest advances in automated data analytics based on machine-learning or statistical analysis will make it possible to extract knowledge out of the data generated at this scale. It is thus critical to consider the numerical experiment as a whole, encompassing both its simulation (HPC) and data-analytics (HPDA) aspects. These two aspects need to be efficiently coupled to overcome the widening performance gap between compute and I/O. This also opens the road for innovative numerical patterns where the outcome of analytics is used to steer the simulation and greatly increase the scientific return of investment for numerical experiments. Workshop |
Thursday, June 24th2:00pm-6:00pmAdvanced MPI Programming details Tutorial Better Scientific Software details Tutorial Compression for Scientific & Engineering Data details Tutorial Determining Parallel Application Execution Efficiency and Scaling using the POP Methodology details Tutorial Getting Started with Containers on HPC details Tutorial Hands-on Practical Hybrid Parallel Application Performance Engineering details Tutorial High Performance Distributed Deep Learning details Tutorial Kokkos: Performance Portability for C++ Applications and Libraries details Tutorial Managing HPC Software Complexity with Spack details Tutorial Mastering Tasking with OpenMP details Tutorial The Scalable Vector Extension: Programming Tools and Performance Analysis details Tutorial | Friday, June 25th2:00pm-6:00pmHands-On HPC Application Development Using C++ and SYCL details Tutorial InfiniBand, High-speed Ethernet, RoCE, Omni-Path, EFA, and Slingshot for Beginners details Tutorial Introduction to HPC: Applications, Systems, and Programming Models details Tutorial Maintaining a Modern Scientific Software Stack Made Easy with EasyBuild details Tutorial Modern Mixed- and Multi-Precision Methods details Tutorial OpenMP Common Core: Learning Parallelization of Real Applications from the Ground-Up details Tutorial Productive Parallel Programming for FPGA with High-Level Synthesis details Tutorial | Friday, July 2nd2:00pm-6:00pm16th Workshop on Virtualization in High-Performance Cloud Computing details Workshop 2nd ISC-HPC International Workshop on “Monitoring and Operational Data Analytics” (MODA) details Workshop 2nd International Workshop on Machine Learning Hardware details Workshop 5th International Workshop on In Situ Visualization details Workshop 7th Annual High Performance Container Workshop details Workshop Approximate and Transprecision Computing on Emerging Technologies (ATCET) - Second Edition details Workshop Arm HPC User Group (AHUG) 2021 details Workshop Compiler-assisted Correctness Checking and Performance Optimization for HPC details Workshop Deep Learning on Supercomputers details Workshop Fourth HPC Applications in Precision Medicine Workshop details Workshop HPC I/O in the Data Center details Workshop ISC'21 SuperCompCloud: 4th International Workshop on Interoperability of Supercomputing and Cloud... details Workshop Machine Learning on HPC Systems details Workshop Numerical Algorithms and Libraries for Exascale details Workshop Second International Workshop on the Application of Machine Learning Techniques to Computational... details Workshop Sixth International Workshop on Communication Architectures for HPC, Big Data, Deep Learning and... details Workshop The Second Workshop on LLVM Compiler and Tools for HPC details Workshop Third Workshop on HPC Education and Training for Emerging Technologies details Workshop Workshop on the In Situ Co-Execution of High-Performance Computing & Data Analysis details Workshop |