Parallel I/O in Low Latency Storage Systems

Date

2022-05-01

Journal Title

Journal ISSN

Volume Title

Publisher

Abstract

The high performance computing (HPC) systems has known a momentous evolution during the last two last decades. Embedding thousands of cores, with very powerful processing capabilities, today’s supercomputers can process a tremendous amount of data in a matter of seconds. However, the evolution of the storage systems has not kept pace with it, which has led to a huge gap between I/O performance and processing performance. Therefore, computer scientists had mainly focused on improving the I/O performance by providing software solutions e.g. collective I/O and asynchronous I/O, that constitute the foundation of parallel I/O. They proposed several algorithms and techniques in order to hide the I/O overhead and improve the overall performance of storage systems by targeting the bandwidth and the capacity. The Message Passing Interface (MPI) has been the most recognized parallel programming paradigm for large scale parallel applications. Starting from version 2 of the MPI specification, the standard has introduced an interface for parallel file I/O support, referred to as MPI-I/O. By extending the MPI concepts to file I/O operations, the programming model becomes more complete and offers more options for developers to exploit the performance benefits of parallel I/O. While reaching the new era of Exascale computing, multiple innovative technologies has risen to the surface opening the door toward a balanced HPC ecosystem that incorporates low latency storage systems. Nevertheless, this evolution has also posed new challenges regarding parallel I/O optimizations. Whereas hardware latency was the main source of I/O overhead, software latency was usually negligible. However, in low latency storage systems, the equation changes since the former would be reduced to the same level as the latter. Therefore, we aim in this dissertation at solving the new equation by providing multiple optimization techniques to the existing parallel I/O solutions within MPI I/O context. In particular, this dissertation targets the communication overhead of collective I/O operations, the computation phase of complex access patterns within independent I/O operations and the file locking overhead of the Lustre parallel file system.

Description

Keywords

HPC, Parallel I/O, MPI

Citation

Portions of this document appear in: Feki, Raafat, and Edgar Gabriel. "On Overlapping Communication and File I/O in Collective Write Operation." In 2020 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW), pp. 1-8. IEEE, 2020; and in: Feki, Raafat, and Edgar Gabriel. "Design and Evaluation of Multi-threaded Optimizations for Individual MPI I/O Operations." In 2022 30th Euromicro International Conference on Parallel, Distributed and Network-based Processing (PDP), pp. 122-126. IEEE, 2022.