In our previous work we have studied the performance of a parallel program, based on a direction splitting approach, solving time dependent Stokes equation. In it, we have used a rectangular uniform mesh, combined with a central difference scheme for the second derivatives. In our work, we were targeting massively parallel computers, as well as clusters of multi-core nodes. Therefore, the developed implementation used hybrid parallelization based on the MPI and OpenMP standards. Specifically, (i) between-node parallelism was supported by using MPI-based communication, while (ii) inside-node parallelism was supported by the OpenMP. In this way, by matching “structure of parallelization” with the architecture of modern large-scale computers, we have attempted at maximizing parallel efficiency of the program.

This paper presents an experimental performance study of the developed parallel implementation on a supercomputer using Intel Xeon processors, as well as Intel Xeon Phi co-processors. The experimental results show an essential improvement when running experiments for a variety of problem sizes and number of cores / threads.

1.
J.-L.
Guermond
and
P.
Minev
(
2010
)
Comptes Rendus Mathematique
348
,
581
585
.
2.
J.-L.
Guermond
and
P.
Minev
(
2011
)
Computer Methods in Applied Mechanics and Engineering
200
,
2083
2093
.
3.
M.
Ganzha
,
K.
Georgiev
,
I.
Lirkov
,
S.
Margenov
, and
M.
Paprzycki
, “
Highly parallel alternating directions algorithm for time dependent problems
,” in
AMiTaNS’11
,
AIP
CP
1404
, edited by
M.D.
Todorov
and
C.I.
Christov
(
American Institute of Physics
,
Melville, NY
,
2011
), pp.
210
217
.
4.
I.
Lirkov
,
M.
Paprzycki
,
M.
Ganzha
, and
P.
Gepner
, “
Performance evaluation of MPI/OpenMP algorithm for 3D time dependent problems
,” in
Preprints of Position Papers of the Federated Conference on Computer Science and Information Systems, Annals of Computer Science and Information Systems
, Vol.
1
, edited by
M.
Ganzha
,
L.
Maciaszek
, and
M.
Paprzycki
(
2013
), pp.
27
32
.
5.
E.
Anderson
,
Z.
Bai
,
C.
Bischof
,
S.
Blackford
,
J.
Demmel
,
J.
Dongarra
,
J. D.
Croz
,
A.
Greenbaum
,
S.
Hammarling
,
A.
McKenney
, and
D.
Sorensen
,
LAPACK Users’ Guide
, 3rd edn (
SIAM
,
Philadelphia
,
1999
).
6.
M.
Ganzha
,
K.
Georgiev
,
I.
Lirkov
, and
M.
Paprzycki
(
2015
)
Computers & Mathematics with Applications
70
,
2762
2772
.
7.
M.
Snir
,
S.
Otto
,
S.
Huss-Lederman
,
D.
Walker
, and
J.
Dongarra
,
MPI: The Complete Reference, Scientific and Engineering Computation Series
(
The MIT Press, Cambridge
,
Massachusetts
,
1997
) second printing.
8.
D.
Walker
and
J.
Dongarra
(
1996
)
Supercomputer
63
,
56
68
.
9.
W.
Gropp
,
E.
Lusk
, and
A.
Skjellum
Using MPI: Portable Parallel Programming with the Message-Passing Interface
(
The MIT Press
,
2014
).
10.
R.
Chandra
,
R.
Menon
,
L.
Dagum
,
D.
Kohr
,
D.
Maydan
, and
J.
McDonald
,
Parallel Programming in OpenMP
(
Morgan Kaufmann
,
2000
).
11.
B.
Chapman
,
G.
Jost
, and
R.
Van Der Pas
,
Using OpenMP: Portable Shared Memory Parallel Programming
, Vol.
10
(
MIT press
,
2008
).
12.
H.
Wang
(
1981
)
ACM Transactions on Mathematical Software (TOMS)
7
,
170
183
.
This content is only available via PDF.
You do not currently have access to this content.