Skip to main content
SHARE
Publication

GPU Data Access on Complex Geometries for D3Q19 Lattice Boltzmann Method

by Gregory Herschlag, Seyong Lee, Jeffrey S Vetter, Amanda Randles
Publication Type
Conference Paper
Book Title
2018 IEEE International Parallel and Distributed Processing Symposium (IPDPS)
Publication Date
Page Numbers
825 to 834
Issue
0
Publisher Location
New Jersey, United States of America
Conference Name
IEEE International Parallel and Distributed Processing Symposium (IPDPS)
Conference Location
Vancouver, Canada
Conference Sponsor
IEEE
Conference Date
-

GPU performance of the lattice Boltzmann method (LBM) depends heavily on memory access patterns. When LBM is advanced with GPUs on complex computational domains, geometric data is typically accessed indirectly, and lattice data is typically accessed lexicographically in the Structure of Array (SoA) layout. Although there are a variety of existing access patterns beyond the typical choices, no study has yet examined the relative efficacy between them. Here, we compare a suite of memory access schemes via empirical testing and performance modeling. We find strong evidence that semi-direct addressing is the superior addressing scheme for the majority of cases examined: Semi-direct addressing increases computational speed and often reduces memory consumption. For lattice layout, we find that the Collected Structure of Arrays (CSoA) layout outperforms the SoA layout. When compared to state-of-the-art practices, our recommended addressing modifications lead to performance gains between 10-40% across different complex geometries, fluid volume fractions, and resolutions. The modifications also lead to a decrease in memory consumption by as much as 17%. Having discovered these improvements, we examine a highly resolved arterial geometry on a leadership class system. On this system we present the first near-optimal strong results for LBM with arterial geometries run on GPUs. We also demonstrate that the above recommendations remain valid for large scale, many device simulations, which leads to an increased computational speed and average memory usage reductions. To understand these observations, we employ performance modeling which reveals that semi-direct methods outperform indirect methods due to a reduced number of total loads/stores in memory, and that CSoA outperforms SoA and bundling due to improved caching behavior.