From: jccharr Date: Wed, 26 Mar 2014 14:06:36 +0000 (+0100) Subject: prediction error X-Git-Tag: ispa14_submission~10 X-Git-Url: https://bilbo.iut-bm.univ-fcomte.fr/and/gitweb/mpi-energy.git/commitdiff_plain/fee64fcb60f40dd77da66ab700cae962ede23ad4?ds=inline prediction error --- diff --git a/paper.tex b/paper.tex index 50ecb90..a89ffb1 100644 --- a/paper.tex +++ b/paper.tex @@ -118,7 +118,6 @@ we conclude in Section~\ref{sec.concl} with a summary and some future works. \section{Related works} \label{sec.relwork} -\AG{Consider introducing the models (sec.~\ref{sec.exe}) before related works} In this section, some heuristics to compute the scaling factor are presented and classified into two categories: offline and online methods. @@ -133,7 +132,7 @@ values could be computed based on information retrieved by analyzing the code of the program and the computing system that will execute it. In ~\cite{40}, Azevedo et al. detect during compilation the dependency points between -tasks in a parallel program. This information is then used to lower the frequency of +tasks in a multi-task program. This information is then used to lower the frequency of some processors in order to eliminate slack times. A slack time is the period of time during which a processor that have already finished its computation, have to wait for a set of processors to finish their computations and send their results to the waiting processor in order to continue its task that is @@ -269,7 +268,6 @@ EQ~(\ref{eq:energy}). The optimal scaling factor is computed by minimizing the d \left( 1 + \sum_{i=2}^{N} \frac{T_i^3}{T_1^3} \right) } \end{equation} -\JC{The following 2 sections can be merged easily} \section{Performance evaluation of MPI programs} \label{sec.mpip} @@ -484,7 +482,7 @@ frequency by the new one see EQ~(\ref{eq:s}). In our cluster there are 18 available frequency states for each processor. This leads to 18 run states for each program. We use seven MPI programs of the NAS parallel benchmarks: CG, MG, EP, FT, BT, LU -and SP. Figure~(\ref{fig:pred}) presents plots of the real execution times and the simulated ones. The maximum normalized error between the predicted execution time and the real time (SimGrid time) for all programs is between 0.0073 to 0.031. The better case is for CG and the worse case is for LU. +and SP. Figure~(\ref{fig:pred}) presents plots of the real execution times and the simulated ones. The maximum normalized error between these two execution times varies between 0.0073 to 0.031 dependent on the executed benchmark. The smallest prediction error was for CG and the worst one was for LU. \subsection{The experimental results for the scaling algorithm } The proposed algorithm was applied to seven MPI programs of the NAS benchmarks (EP, CG, MG, FT, BT, LU and SP) which were run with three classes (A, B and