Skip to Main content Skip to Navigation
Journal articles

Efficient GPU Implementation of the Linearly Interpolated Bounce-Back Boundary Condition

Abstract : Interpolated bounce-back boundary conditions for the lattice Boltzmann method (LBM) make the accurate representation of complex geometries possible. In the present work, we describe an implementation of a linearly interpolated bounce-back (LIBB) boundary condition for graphics processing units (GPUs). To validate our code, we simulated the flow past a sphere in a square channel. At low Reynolds numbers, results are in good agreement with experimental data. Moreover, we give an estimate of the critical Reynolds number for transition from steady to periodic flow. Performance recorded on a single node server with eight GPU based computing devices ranged up to 2.63×1092.63×109 fluid node updates per second. Comparison with a simple bounce-back version of the solver shows that the impact of LIBB on performance is fairly low.
Complete list of metadatas

Cited literature [16 references]  Display  Hide  Download
Contributor : Laboratoire Cethil <>
Submitted on : Monday, June 9, 2014 - 2:49:03 PM
Last modification on : Tuesday, October 13, 2020 - 2:50:05 PM
Long-term archiving on: : Tuesday, September 9, 2014 - 10:38:07 AM


Files produced by the author(s)




C. Obrecht, F. Kuznik, Bernard Tourancheau, J.-J. Roux. Efficient GPU Implementation of the Linearly Interpolated Bounce-Back Boundary Condition. Computers and Mathematics with Applications, Elsevier, 2013, 65 (6), ⟨10.1016/j.camwa.2012.05.014⟩. ⟨hal-00731150⟩



Record views


Files downloads