-\title{Energy Consumption Reduction with DVFS for Message \\
+\title{Optimizing Energy Consumption with DVFS for Message \\
Passing Iterative Applications on \\
- Grid Architecture}
+ Grid Architectures}
different dynamic and static powers from the nodes of the other clusters,
noted as $\Pd[ij]$ and $\Ps[ij]$ respectively. Therefore, even if the distributed
message passing iterative application is load balanced, the computation time of each CPU $j$
-in cluster $i$ noted $\Tcp[ij]$ may be different and different frequency scaling factors may be
+in cluster $i$ noted $\Tcp[ij]$ may be slightly different due to the delay caused by the scheduler of the operating system. Therefore, different frequency scaling factors may be
computed in order to decrease the overall energy consumption of the application
-and reduce slack times. The communication time of a processor $j$ in cluster $i$ is noted as
+and reduce the slack times. The communication time of a processor $j$ in cluster $i$ is noted as
$\Tcm[ij]$ and could contain slack times when communicating with slower nodes,
see Figure~\ref{fig:heter}. Therefore, all nodes do not have equal
communication times. While the dynamic energy is computed according to the
\begin{figure}[!t]
\centering
\includegraphics[scale=0.6]{fig/power_consumption.pdf}
- \AG{I don't understand the labels on the horizontal axis: 10:30:37, 10:30:38,
- etc.}
\caption{The power consumption by one core from the Taurus cluster}
\label{fig:power_cons}
\end{figure}
\begin{tabular}{|*{7}{c|}}
\hline
& & Max & Min & Diff. & & \\
- Cluster & CPU & Freq. & Freq. & Freq. & No. of cores & Dynamic power \\
+ Cluster & CPU & Freq. & Freq. & Freq. & Cores & Dynamic power \\
Name & model & GHz & GHz & GHz & per CPU & of one core \\
\hline
& Intel & & & & & \\
\begin{tabular}{|*{4}{c|}}
\hline
\multirow{2}{*}{Scenario name} & \multicolumn{3}{c|} {The participating clusters} \\ \cline{2-4}
- & Cluster & Site & No. of nodes \\
+ & Cluster & Site & Nodes per cluster \\
\hline
\multirow{3}{*}{Two sites / 16 nodes} & Taurus & Lyon & 5 \\ \cline{2-4}
& Graphene & Nancy & 5 \\ \cline{2-4}
The execution times of these benchmarks
over one site with 16 and 32 nodes are also lower when compared to those of the two sites
-scenario. Moreover, most of the benchmarks running over the one site scenario their execution times are approximately divided by two when the number of computing nodes is doubled from 16 to 32 nodes (linear speed up according to the number of the nodes).\AG{Parse error (cannot understand the previous sentence).}
+scenario. Moreover, most of the benchmarks running over the one site scenario have their execution times approximately divided by two when the number of computing nodes is doubled from 16 to 32 nodes (linear speed up according to the number of the nodes).
However, the execution times and the energy consumptions of EP and MG
benchmarks, which have no or small communications, are not significantly
\caption{The multicores scenarios}
\begin{tabular}{|*{4}{c|}}
\hline
-Scenario name & Cluster name & \begin{tabular}[c]{@{}c@{}}No. of nodes\\ in each cluster\end{tabular} &
- \begin{tabular}[c]{@{}c@{}}No. of cores\\ for each node\end{tabular} \\ \hline
+Scenario name & Cluster name & Nodes per cluster &
+ Cores per node \\ \hline
\multirow{3}{*}{One core per node} & Graphite & 4 & 1 \\ \cline{2-4}
& Graphene & 14 & 1 \\ \cline{2-4}
& Griffon & 14 & 1 \\ \hline