-The overall energy consumption was computed for each instance according to the
-energy consumption model (\ref{eq:energy}), with and without applying the
-algorithm. The execution time was also measured for all these experiments. Then,
-the energy saving and performance degradation percentages were computed for each
-instance. The results are presented in Tables~\ref{table:res_4n},
-\ref{table:res_8n}, \ref{table:res_16n}, \ref{table:res_32n},
-\ref{table:res_64n} and \ref{table:res_128n}. All these results are the average
-values from many experiments for energy savings and performance degradation.
-The tables show the experimental results for running the NAS parallel benchmarks
-on different number of nodes. The experiments show that the algorithm
-significantly reduces the energy consumption (up to \np[\%]{35}) and tries to
-limit the performance degradation. They also show that the energy saving
-percentage decreases when the number of computing nodes increases. This
-reduction is due to the increase of the communication times compared to the
-execution times when the benchmarks are run over a higher number of nodes.
-Indeed, the benchmarks with the same class, C, are executed on different numbers
-of nodes, so the computation required for each iteration is divided by the
-number of computing nodes. On the other hand, more communications are required
-when increasing the number of nodes so the static energy increases linearly
-according to the communication time and the dynamic power is less relevant in
-the overall energy consumption. Therefore, reducing the frequency with
-Algorithm~\ref{HSA} is less effective in reducing the overall energy savings. It
-can also be noticed that for the benchmarks EP and SP that contain little or no
-communications, the energy savings are not significantly affected by the high
-number of nodes. No experiments were conducted using bigger classes than D,
-because they require a lot of memory (more than \np[GB]{64}) when being executed
-by the simulator on one machine. The maximum distance between the normalized
-energy curve and the normalized performance for each instance is also shown in
-the result tables. It decrease in the same way as the energy saving percentage.
-The tables also show that the performance degradation percentage is not
-significantly increased when the number of computing nodes is increased because
-the computation times are small when compared to the communication times.
-
-Figures~\ref{fig:energy} and \ref{fig:per_deg} present the energy saving and
-performance degradation respectively for all the benchmarks according to the
-number of used nodes. As shown in the first plot, the energy saving percentages
-of the benchmarks MG, LU, BT and FT decrease linearly when the number of nodes
-increase. While for the EP and SP benchmarks, the energy saving percentage is
-not affected by the increase of the number of computing nodes, because in these
-benchmarks there are little or no communications. Finally, the energy saving of
-the GC benchmark significantly decrease when the number of nodes increase
-because this benchmark has more communications than the others. The second plot
-shows that the performance degradation percentages of most of the benchmarks
-decrease when they run on a big number of nodes because they spend more time
-communicating than computing, thus, scaling down the frequencies of some nodes
-has less effect on the performance.
-
-\subsection{The results for different power consumption scenarios}
-\label{sec.compare}
-
-The results of the previous section were obtained while using processors that
-consume during computation an overall power which is \np[\%]{80} composed of
-dynamic power and of \np[\%]{20} of static power. In this section, these ratios
-are changed and two new power scenarios are considered in order to evaluate how
-the proposed algorithm adapts itself according to the static and dynamic power
-values. The two new power scenarios are the following:
-
-\begin{itemize}
-\item \np[\%]{70} of dynamic power and \np[\%]{30} of static power
-\item \np[\%]{90} of dynamic power and \np[\%]{10} of static power
-\end{itemize}
-
-The NAS parallel benchmarks were executed again over processors that follow the
-new power scenarios. The class C of each benchmark was run over 8 or 9 nodes
-and the results are presented in Tables~\ref{table:res_s1} and
-\ref{table:res_s2}. These tables show that the energy saving percentage of the
-\np[\%]{70}-\np[\%]{30} scenario is smaller for all benchmarks compared to the
-energy saving of the \np[\%]{90}-\np[\%]{10} scenario. Indeed, in the latter
-more dynamic power is consumed when nodes are running on their maximum
-frequencies, thus, scaling down the frequency of the nodes results in higher
-energy savings than in the \np[\%]{70}-\np[\%]{30} scenario. On the other hand,
-the performance degradation percentage is smaller in the \np[\%]{70}-\np[\%]{30}
-scenario compared to the \np[\%]{90}-\np[\%]{10} scenario. This is due to the
-higher static power percentage in the first scenario which makes it more
-relevant in the overall consumed energy. Indeed, the static energy is related
-to the execution time and if the performance is degraded the amount of consumed
-static energy directly increases. Therefore, the proposed algorithm does not
-really significantly scale down much the frequencies of the nodes in order to
-limit the increase of the execution time and thus limiting the effect of the
-consumed static energy.
-
-Both new power scenarios are compared to the old one in
-Figure~\ref{fig:sen_comp}. It shows the average of the performance degradation,
-the energy saving and the distances for all NAS benchmarks of class C running on
-8 or 9 nodes. The comparison shows that the energy saving ratio is proportional
-to the dynamic power ratio: it is increased when applying the
-\np[\%]{90}-\np[\%]{10} scenario because at maximum frequency the dynamic energy
-is the most relevant in the overall consumed energy and can be reduced by
-lowering the frequency of some processors. On the other hand, the energy saving
-decreases when the \np[\%]{70}-\np[\%]{30} scenario is used because the dynamic
-energy is less relevant in the overall consumed energy and lowering the
-frequency does not return big energy savings. Moreover, the average of the
-performance degradation is decreased when using a higher ratio for static power
-(e.g. \np[\%]{70}-\np[\%]{30} scenario and \np[\%]{80}-\np[\%]{20}
-scenario). Since the proposed algorithm optimizes the energy consumption when
-using a higher ratio for dynamic power the algorithm selects bigger frequency
-scaling factors that result in more energy saving but less performance, for
-example see Figure~\ref{fig:scales_comp}. The opposite happens when using a
-higher ratio for static power, the algorithm proportionally selects smaller
-scaling values which result in less energy saving but also less performance
-degradation.
-
-\begin{table}[!t]
- \caption{The results of the \np[\%]{70}-\np[\%]{30} power scenario}
- % title of Table
- \centering
- \begin{tabular}{|*{6}{r|}}
- \hline
- Program & Energy & Energy & Performance & Distance \\
- name & consumption/J & saving\% & degradation\% & \\
- \hline
- CG & 4144.21 & 22.42 & 7.72 & 14.70 \\
- \hline
- MG & 1133.23 & 24.50 & 5.34 & 19.16 \\
- \hline
- EP & 6170.30 & 16.19 & 0.02 & 16.17 \\
- \hline
- LU & 39477.28 & 20.43 & 0.07 & 20.36 \\
- \hline
- BT & 26169.55 & 25.34 & 6.62 & 18.71 \\
- \hline
- SP & 19620.09 & 19.32 & 3.66 & 15.66 \\
- \hline
- FT & 6094.07 & 23.17 & 0.36 & 22.81 \\
- \hline
- \end{tabular}
- \label{table:res_s1}
-\end{table}