IWOMP 2020
IWOMP 2020 – Conference Program
Due to the COVID-19 pandemic, the IWOMP 2020 workshop will be completely virtual. Tutorial Day will take place on Monday, Sept. 21,
followed by the technical program, Sept. 22-24. To help accommodate EU participation the program was changed to three 6-hr days. The Tutorial Day will, however, be 7 hours in length. All times are given for US Central Daylight Time (CDT), UTC -5.
Presentations Slides
The presentation slides for each presentation can be found below each entry in the program, where available.
Conference Proceedings
Free access to the Springer proceedings will be available to conference participants for 4 weeks, (Sep 22 – Oct 19). To gain access to the proceedings participants should follow the link below and use the password that will be provided to them during the online presentations.
- Order conference proceedings – Springer eBook and Softcover (free worldwide delivery)
- View conference proceedings (Delegates only)
Monday, September 21 | Tutorials
All times are shown in US Central Daylight Time (CDT) zone.
Time | Tutorial Name | Presenters |
---|---|---|
8:00 – 10:00 | OpenMP Common Core
In this tutorial, we’ll undertake an interactive exploration of the common core of OpenMP. We will cover through discussion and demos the 20 features of OpenMP programmers use all the time. This will give you what you need to become an OpenMP programmer. It will also prepare you for the more advanced OpenMP tutorials later in the day.The tutorial will be based on the C programming language, though Fortran programmers should be able to keep up as we avoid the more complex corners of C. The contents of the tutorial focusses on novice OpenMP programmers, but more experienced OpenMP programmers will benefit as well; they’ll just need to download the exercises (I’ll provide a link to a GitHub repository) so they can explore on their own during the tutorial, adding all the complexity they need to stay engaged. |
Tim Mattson, Intel![]() |
10:00 – 10:30 | Break | |
10:30 – 12:30 | Advanced OpenMP
This tutorial will cover several advanced topics of OpenMP:
|
Michael Klemm, Intel & Christian Terboven, RWTH Aachen University ![]() organization of the Intel Architecture, Graphics and Software group. His focus is on High Performance and Throughput Computing. He holds an M.Sc. in Computer Science and a Doctor of Engineering degree (Dr.-Ing.) in Computer Science from the Friedrich-Alexander-University Erlangen-Nuremberg, Germany. Michael’s research focus is on compilers and runtime optimizations for distributed systems. His areas of interest include compiler construction, design of programming languages, parallel programming, and performance analysis and tuning. In 2016, Michael was appointed Chief Executive Officer of the OpenMP Architecture Review Board.
|
12:20 – 13:00 | Break | |
13:00 – 15:00 | OpenMP Offload (GPUs)
While most HPC developers have MPI experience, many are not familiar with the latest OpenMP features and how to use them in their codes. Modern OpenMP can be used in creative ways to map parallelism to current and emerging parallel architectures. This tutorial that describes recent and new features of OpenMP for accelerators, with a focus on those features that have proven important applications.The OpenMP 5.1 specification is scheduled for release at SC’20 with exciting new functionality that improves support for Accelerators.Thus it is important not only for developers to be aware of the current standard, and what is available today, but also what is coming next and what will be available in the exascale time frame (2021).The tutorial is expected to cover the following topics:
|
Oscar Hernandez, Oak Ridge National Laboratory & Tom Scogland, Lawrence Livermore National Laboratory ![]()
|
Tuesday, September 22 | Program
All times are shown in US Central Daylight Time (CDT) zone. Presentation Slides are provided at the discretion of the paper authors,
and will not be available until the presentation or shortly afterwards.
Time | Presentation |
---|---|
08:00 | Introduction
Dan Stanzione, Director of TACC (UT, Texas) |
08:10 | Keynote I: A Tale of Four Packages
Jack Dongarra, Distinguished Professor, University of Tennessee, Oak Ridge National Laboratory and University of Manchester Abstract. In this talk we will look at four software packages for dense linear algebra. These packages have been developed over time. Some of these packages are intended as research vehicles and others as production tools. The four packages are: BALLISTIC to sustain LAPACK and ScaLAPACK packages. PLASMA research into multicore implementations. MAGMA research into GPU implementations. SLATE dense linear algebra for Exascale machines. We will look at how OpenMP is being used in each.
He is a Fellow of the AAAS, ACM, IEEE, and SIAM and a foreign member of the Russian Academy of Science, a foreign member of the British Royal Society, and a member of the US National Academy of Engineering. |
09:00 | Papers I: Performance Methodologies |
BEST PAPER
FAROS: A Framework To Analyze OpenMP Compilation Through Benchmarking and Compiler Optimization Analysis Giorgis Georgakoudis, Johannes Doerfert, Ignacio Laguna and Tom Scogland |
|
Evaluating the Efficiency of OpenMP Tasking for Unbalanced Computation on Diverse CPU Architectures
Stephen L. Olivier |
|
10:00 | Break |
10:30 | Papers II: Applications |
A Case Study of Porting HPGMG from CUDA to OpenMP Target Offload
Christopher Daley, Hadia Ahmed, Samuel Williams and Nicholas Wright |
|
P-Aevol: an OpenMP Parallelization of a Biological Evolution Simulator, Through Decomposition in Multiple Loops
Laurent Turpin, Thierry Gautier, Jonathan Rouzaud-Cornabas and Christian Perez |
|
Evaluating Performance of OpenMP Tasks in a Seismic Stencil Application
Eric Raut, Jie Meng, Mauricio Araya-Polo and Barbara Chapman |
|
12:00 | Break |
12:30 | Papers III: OpenMP Extensions |
Unified Sequential Optimization Directives in OpenMP
Brandon Neth, Tom Scogland, Michelle Mills Strout and Bronis R. de Supinski |
|
Support Data Shuffle Between Threads in OpenMP
Anjia Wang, Xinyao Yi and Yonghong Yan |
|
13:30 | OpenMP in MLIR and Flang
ARM Sponsor Tech Talk. David Truby, Arm. Abstract. Flang is a new Fortran frontend in the LLVM project. While Flang is a work in progress, when finished it will provide comprehensive support for the latest Fortran and OpenMP standards. Flang uses the new LLVM MLIR technology to generate Fortran and OpenMP codes. This talk will give an overview of status and describe how OpenMP support is being implemented in Flang and MLIR.Flang is a new Fortran frontend in the LLVM project. While Flang is a work in progress, when finished it will provide comprehensive support for the latest Fortran and OpenMP standards. Flang uses the new LLVM MLIR technology to generate Fortran and OpenMP codes. This talk will give an overview of status and describe how OpenMP support is being implemented in Flang and MLIR.
Speaker Profile. David Truby is a senior software engineer at Arm. His latest project has been adding OpenMP support for the LLVM Flang Fortran compiler. He is a PhD candidate at the University of Warwick, with his research focusing on performance portability frameworks and parallel programming models, and has previously worked with IBM Research on OpenMP target offload support in the Clang compiler. |
Wednesday, September 23 | Program
All times are shown in US Central Daylight Time (CDT) zone. Presentation Slides are provided at the discretion of the paper authors,
and will not be available until the presentation or shortly afterwards.
Time | Presentation |
---|---|
08:00 | Keynote II: Programming Models for a Mixed-Signal AI Inference Accelerator
Eric Stotzer, Mythic Inc. Abstract. This talk will cover Mythic’s hybrid mixed-signal computing architecture and unique software development tools, including a Deep Neural Network (DNN) graph compiler. In addition, some ideas will be proposed on how OpenMP can be used to program this type of architecture. Mythic’s Intelligence Processing Units (IPUs) combine analog compute-in-memory acceleration with digital processing elements. They are designed for high-performance and power-efficient AI inference acceleration.
The Mythic IPU is a tile-based dataflow architecture. Each tile has an analog compute array, flash memory for weight storage, local SRAM memory, a single-instruction multiple-data (SIMD) unit, and a control processor. The tiles are interconnected with an efficient on-chip router network. Mythic has built a unique suite of development tools, including a DNN graph compiler, to enable the rapid deployment of AI inference applications on the IPU. The tools perform such actions as mapping DNNs to tiles, setting up dataflow conditions, and analog-aware program transformations.
|
09:00 | Papers IV: Performance Studies |
Performance Study of SpMV Towards an Auto-tuned and Task-based SpMV (LASs Library)
Sandra Catalan, Tetsuzo Usui, Leonel Toledo, Xavier Martorell, Jesús Labarta and Pedro Valero-Lara |
|
A Case Study on Addressing Complex Load Imbalance in OpenMP
Fabian Orland and Christian Terboven |
|
10:00 | Break |
10:30 | Papers V: Tools |
On-the-fly Data Race Detection with the Enhanced OpenMP Series-Parallel Graph
Nader Boushehrinejadmoradi, Adarsh Yoga and Santosh Nagarakatte |
|
AfterOMPT: An OMPT-based tool for Fine-Grained Tracing of Tasks and Loops
Igor Wodiany, Andi Drebes, Richard Neill and Antoniu Pop |
|
Co-designing OpenMP Programming Model Features with OMPT and Simulation
Matthew Baker, Oscar Hernandez and Jeffrey Young |
|
12:00 | Break |
12:30 | Papers VI: NUMA |
sOMP: Simulating OpenMP Task-based Applications with NUMA Effects
Idriss Daoudi, Philippe Virouleau, Thierry Gautier, Samuel Thibault and Olivier Aumage |
|
Virtflex: Automatic Adaptation to NUMA Topology Change for OpenMP Applications
Runhua Zhang, Alan L. Cox and Scott Rixner |
|
13:30 | Intel oneAPI Toolkits
Intel Sponsor Tech Talk. Ron Green, Intel. |
Thursday, September 24 | Program
All times are shown in US Central Daylight Time (CDT) zone. Presentation Slides are provided at the discretion of the paper authors,
and will not be available until the presentation or shortly afterwards.
Time | Presentation |
---|---|
08:00 | Keynote III: OpenMP Enables the Next Frontier of High-Performance Computing Mark Papermaster, CTO & EVP of Technology and Engineering, AMD Abstract. In his keynote, Mark will discuss AMD’s commitment to High Performance Computing, an open ecosystem and how OpenMP fits squarely in AMD’s product strategy to enable heterogeneous computing. He will also take a closer look at AMD’s approach to optimize for performance while maintaining compatibility with multiple programming languages, and heterogeneous architectures. He will also touch on the expected impact for Machine Learning and future Exascale supercomputers. .
Mark Papermaster is Chief Technology Officer and Executive Vice President of Technology and Engineering at AMD and is responsible for corporate technical direction, product development including system-on-chip (SOC) methodology, microprocessor design, I/O and memory and advanced research. He led the re-design of engineering processes at AMD and the development of the award-winning “Zen” high-performance x86 CPU family, high-performance GPUs and the company’s modular design approach, Infinity Fabric. He also oversees Information Technology that delivers AMD’s compute infrastructure and services. His more than 35 years of engineering experience includes significant leadership roles managing the development of a wide range of products, from microprocessors to mobile devices and high-performance servers. Before joining AMD in October 2011 as Chief Technology Officer and Senior Vice President, Papermaster was the leader of Cisco’s Silicon Engineering Group, the organization responsible for silicon strategy, architecture, and development for the company’s switching and routing businesses. He served as Apple Senior Vice President of Devices Hardware Engineering, where he was responsible for iPod and iPhone hardware development. He also held a number of senior leadership positions at IBM overseeing development of the company’s key microprocessor and server technologies. Papermaster received his bachelor’s degree from the University of Texas at Austin and master’s degree from the University of Vermont, both in Electrical Engineering. He is a long-term member of the University of Texas Cockrell School of Engineering Advisory Board, Olin College Presidents Council and the Juvenile Diabetes Research Foundation. Most recently, he was appointed to the CTO Forum Advisory Board and IEEE Industry Advisory Board. |
09:00 | Papers VII: Compilation Techniques |
Using OpenMP to Detect and Speculate Dynamic DOALL Loops
Bruno Chinelato Honorio, João P. L. de Carvalho, Munir Skaf and Guido Araujo |
|
ComPar: Optimized Multi-Compiler for Automatic OpenMP S2S Parallelization
Idan Mosseri, Lee-Or Alon, Re’Em Harel and Gal Oren |
|
10:00 | Break |
10:30 | Papers VIII: Heterogeneous Computing |
OpenMP Device Offloading to FPGAs Using the Nymble Infrastructure
Jens Huthmann, Lukas Sommer, Artur Podobas, Andreas Koch and Kentaro Sano |
|
Data Transfer and Reuse Analysis Tool for GPU-offloading Using OpenMP
Alok Mishra, Abid Malik and Barbara Chapman |
|
Toward Supporting Multi-GPU Targets via Taskloop and User-defined Schedules
Vivek Kale, Wenbin Lu, Anthony Curtis, Abid Malik, Barbara Chapman and Oscar Hernandez |
|
12:00 | Break |
12:30 | Papers IX: Memory |
Preliminary Experience with OpenMP Management Implementation Memory
Adrien Roussel, Patrick Carribault and Julien Jaeger |
|
Memory Anomalies in OpenMP
Lechen Yu, Joachim Protze, Oscar Hernandez and Vivek Sarkar |
|
13:30 | The ROCm Software Stack AMD Sponsor Tech Talk. Gregory Rodgers, AMD.Presentation Slides |