Lattice Boltzmann methods on the ClearSpeed Advance™ accelerator board

  • V. Heuveline
  • J.-P. WeißEmail author


Numerical analysts and programmers are currently facing a conceptual change in processor technology. Multicore concepts, coprocessors and accelerators are becoming a vital part in scientific computing. The new hardware techno- logies lead to new paradigms and require adapted methodologies and techniques in numerical simulation. These developments play an important role in computational fluid dynamics (CFD) where many highly CPU-time demanding problems arise. In this paper, we propose a parallel lattice Boltzmann method (LBM) in the context of a coprocessor technology, the ClearSpeed Advance™ accelerator board. Implementations of LBMs on parallel architectures benefit from localities of the necessary interactions and the regular structure of the underlying meshes. The considered board supports high-level parallelism and double precision conforming to the IEEE 754 standard. However, the solution process relies on a huge amount of data which needs to propagate along the mesh. This prototypical fact shows up the bottleneck of internal communication bandwidth and indicates the limits of this type of small-scale parallel systems.


European Physical Journal Special Topic Lattice Boltzmann Method Double Precision Propagation Step Single Instruction Multiple Data 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. M.C. Sukop, D.T. Thorne, Lattice Boltzmann Modeling (Springer, Berlin, 2006)Google Scholar
  2. B. Chopard, M. Droz, Cellular Automata Modeling of Physical Systems (Cambridge University Press, Cambridge, 1998)Google Scholar
  3. J.P. Weiß, Ph.D. thesis, University of Karlsruhe, 2006Google Scholar
  4. C. Körner, T. Pohl, U. Rüde, N. Thürey, T. Zeiser, Lecture Notes in Computational Science and Engineering 51 (Springer, 2006), chap. 13, pp. 439–466Google Scholar
  5. T. Pohl, F. Deserno, N. Thürey, U. Rüde, P. Lammers, G. Wellein, T. Zeiser, in Proceedings of the ACM/IEEE Conference on Supercomputing (2004), p. 21Google Scholar
  6. J.D. Owens, D. Luebke, N. Govindaraju, M. Harris, J. Krüger, A.E. Lefohn, T.J. Purcell, Computer Graphics Forum 26, 80 (2007)Google Scholar
  7. D. Göddeke, R. Strzodka, S. Turek, in Simulationstechnique 18th Symposium in Erlangen (2005), pp. 139–144Google Scholar
  8. M. deLorimier, A. DeHon, in Proceedings of the ACM/SIGDA 13th International Symposium on Field-Programmable Gate Arrays (2005), pp. 75–85Google Scholar
  9. J. Kurzak, J. Dongarra, Concurrency and Computation: Practice and Experience 19, 1371 (2007)Google Scholar
  10. http://www.clearspeed.comGoogle Scholar
  11. Scholar
  12. Scholar
  13. S. Chen, G.D. Doolen, Ann. Rev. Fluid. Mech. 30, 329 (1998)Google Scholar
  14. X. He, L.S. Luo, J. Stat. Phys. 88, 927 (1997)Google Scholar
  15. J.L. Guermond, C. Migeon, G. Pineau, L. Quartapelle, J. Fluid Mech. 450, 169 (2002)Google Scholar
  16. K. Mattila, J. Hyvaluoma, T. Rossi, M. Aspnas, J. Westerholm, Comp. Phys. Comm. 176, 200 (2007)Google Scholar
  17. V. Heuveline, J. Latt, Int. J. Mod. Phys. C 4, 627 (2007)Google Scholar
  18. http://openlb.orgGoogle Scholar
  19. G. Wellein, T. Zeiser, G. Hager, S. Donath, Comput. Fluids 35, 910 (2006)Google Scholar

Copyright information

© EDP Sciences and Springer 2009

Authors and Affiliations

  1. 1.Institute for Applied and Numerical Mathematics, Karlsruhe Institute of Technology (KIT), Universität Karlsruhe (TH)KarlsruheGermany
  2. 2.SRG New Frontiers in High Performance Computing Exploiting Multicore and Coprocessor Technology, Karlsruhe Institute of Technology (KIT), Universität Karlsruhe (TH)KarlsruheGermany

Personalised recommendations