of asynchronous iterative algorithms. For that, we compare the behaviour of a
synchronous GMRES algorithm with an asynchronous multisplitting one with
simulations which let us easily choose some parameters. Both codes are real MPI
of asynchronous iterative algorithms. For that, we compare the behaviour of a
synchronous GMRES algorithm with an asynchronous multisplitting one with
simulations which let us easily choose some parameters. Both codes are real MPI
$X_{0}$ to find an approximate value $X^*$ of the solution with a very low residual error. Several well-known methods
demonstrate the convergence of these algorithms~\cite{BT89,Bahi07}.
$X_{0}$ to find an approximate value $X^*$ of the solution with a very low residual error. Several well-known methods
demonstrate the convergence of these algorithms~\cite{BT89,Bahi07}.
into several \emph{blocks} that will be solved in parallel on multiple
processing units. The latter will communicate each intermediate results before a
new iteration starts and until the approximate solution is reached. These
into several \emph{blocks} that will be solved in parallel on multiple
processing units. The latter will communicate each intermediate results before a
new iteration starts and until the approximate solution is reached. These
convergence depends on the delay of messages. With synchronous iterations, the
number of iterations is exactly the same than in the sequential mode (if the
parallelization process does not change the algorithm). So the difficulty with
convergence depends on the delay of messages. With synchronous iterations, the
number of iterations is exactly the same than in the sequential mode (if the
parallelization process does not change the algorithm). So the difficulty with
with real data. In fact, from an execution to another the order of messages will
change and the number of iterations to reach the convergence will also change.
According to all the parameters of the platform (number of nodes, power of
with real data. In fact, from an execution to another the order of messages will
change and the number of iterations to reach the convergence will also change.
According to all the parameters of the platform (number of nodes, power of
%Then we have changed the network configuration using three clusters containing
%respectively 33, 33 and 34 hosts, or again by on hundred hosts for all the
%clusters. In the same way as above, a judicious choice of key parameters has
%permitted to get the results in Table~\ref{tab.cluster.3x33} which shows the
%relative gains greater than 1 with a matrix size from 62 to 100 elements.
%Then we have changed the network configuration using three clusters containing
%respectively 33, 33 and 34 hosts, or again by on hundred hosts for all the
%clusters. In the same way as above, a judicious choice of key parameters has
%permitted to get the results in Table~\ref{tab.cluster.3x33} which shows the
%relative gains greater than 1 with a matrix size from 62 to 100 elements.
-\CER{En accord avec RC, on a pour le moment enlevé les tableaux 2 et 3 sachant que les résultats obtenus sont limites. De même, on a enlevé aussi les deux dernières colonnes du tableau I en attendant une meilleure performance et une meilleure precision}
+%\CER{En accord avec RC, on a pour le moment enlevé les tableaux 2 et 3 sachant que les résultats obtenus sont limites. De même, on a enlevé aussi les deux dernières colonnes du tableau I en attendant une meilleure performance et une meilleure precision}
- \item Intra-cluster network bandwidth: \np[Gbit/s]{1.25} and latency: \np[$\mu$s]{0.05};
- \item Inter-cluster network bandwidth: \np[Mbit/s]{5} or \np[Mbit/s]{50} and latency: \np[$\mu$s]{20};
+ \item Intra-cluster network bandwidth: \np[Gbit/s]{1.25} and latency: \np[$\mu$s]{50};
+ \item Inter-cluster network bandwidth: \np[Mbit/s]{5} or \np[Mbit/s]{50} and latency: \np[ms]{20};