X-Git-Url: https://bilbo.iut-bm.univ-fcomte.fr/and/gitweb/mpi-energy2.git/blobdiff_plain/64ae1ea15de81e7d26e31750c558df497b58ab44..7dac8c70ce35d77f8dc3dfc2fc05a2a750dee186:/Heter_paper.tex?ds=inline diff --git a/Heter_paper.tex b/Heter_paper.tex index 690c6a1..e8c5860 100644 --- a/Heter_paper.tex +++ b/Heter_paper.tex @@ -82,7 +82,7 @@ \section{Introduction} \label{sec.intro} Modern processors continue to increased in a performance, achieved maximum number of floating point operations per second (FLOPS), thus the energy consumption and the heat dissipation are increased drastically according to this increase. The number of FLOPS is linearly related to power consumption of a CPU~\cite{51}. -As an example of more power hungry cluster, according to the Top500 list in June 2014 \cite{43}, Tianhe-2 has more than 3 millions of cores and consumed more than 17.8 megawatt per second. Moreover, according to the U.S. annual energy outlook 2014 \cite{60}, the price of energy for 1 megawatt per hour is approximately equal to 70\$ (1.16\$ for megawatt per second). Therefore, we can consider the price of the energy consumption for the Tianhe-2 platform is approximately more than 390 millions dollars of megawatt per year. For this reason, the heterogeneous clusters must be offer more energy efficiency due to the increase in the energy cost and the environment influences. Therefore, a green computing clusters are require nowadays. For example, the GSIC center of Tokyo heterogeneous cluster became the top of the Green500 list in June 2014 \cite{59}. This platform has more than four thousand of MFLOPS per watt. Dynamic voltage and frequency scaling (DVFS) is a process used widely to reduce the energy consumption of the processor. In a heterogeneous clusters enabled DVFS, many researchers used DVFS in a different ways. DVFS can be minimized the energy consumption but it lead to a disadvantage due to the performance degradation increase. Therefore, researchers used different optimization strategies to overcame this problem. The best tradeoff relation between the energy reduction and performance degradation ratio is become a key challenges in a heterogeneous platforms. In this paper we are propose a heterogeneous scaling algorithm that selects the optimal vector of the frequency scaling factors for distributed iterative application, producing minimum energy saving against minimum performance degradation ratio simultaneously. The algorithm has very small overhead, works online and not needs for any training or profiling. +As an example of more power hungry cluster, according to the Top500 list in June 2014 \cite{43}, Tianhe-2 has more than 3 millions of cores and consumed more than 17.8 megawatt per second. Moreover, according to the U.S. annual energy outlook 2014 \cite{60}, the price of energy for 1 megawatt per hour is approximately equal to 70\$ (1.16\$ for megawatt per second). Therefore, we can consider the price of the energy consumption for the Tianhe-2 platform is approximately more than 390 millions dollars of megawatt per year. For this reason, the heterogeneous clusters must be offer more energy efficiency due to the increase in the energy cost and the environment influences. Therefore, a green computing clusters with maximum number of FLOPS per watt are required nowadays. For example, the GSIC center of Tokyo heterogeneous cluster became the top of the Green500 list in June 2014 \cite{59}. This platform has more than four thousand of MFLOPS per watt. Dynamic voltage and frequency scaling (DVFS) is a process used widely to reduce the energy consumption of the processor. In a heterogeneous clusters enabled DVFS, many researchers used DVFS in a different ways. DVFS can be minimized the energy consumption but it leads to a disadvantage due to increase in performance degradation. Therefore, researchers used different optimization strategies to overcame this problem. The best tradeoff relation between the energy reduction and performance degradation ratio is became a key challenges in a heterogeneous platforms. In this paper we are propose a heterogeneous scaling algorithm that selects the optimal vector of the frequency scaling factors for distributed iterative application, producing minimum energy saving against minimum performance degradation ratio simultaneously. The algorithm has very small overhead, works online and not needs for any training or profiling. This paper is organized as follows: Section~\ref{sec.relwork} presents some related works from other authors. Section~\ref{sec.exe} describes how the @@ -660,8 +660,10 @@ The maximum distance between the normalized energy curve and the normalized perf \end{figure} - \textbf{ The energy saving and performance degradation of all benchmarks are plotted to the number of -nodes as in plots (\ref{fig:energy} and \ref{fig:per_deg}). As shown in the plots, the energy saving percentage of the benchmarks MG, LU, BT and FT is decreased linearly when the the number of nodes increased. While in EP benchmark the energy saving percentage is approximately the same percentage when the number of computing nodes is increased, because in this benchmark there is no communications. In the SP benchmark the energy saving percentage is decreased when it runs on a small number of nodes, while this percentage is increased when it runs on a big number of nodes. The energy saving of the GC benchmarks is significantly decreased when the number of nodes is increased, because this benchmark has more communications compared to other benchmarks. The performance degradation percentage of the benchmarks CG, EP, LU and BT is decreased when they run on a big number of nodes. While in MG benchmark has a higher percentage of performance degradation when it runs on a big number of nodes. The inverse happen in SP benchmark has smaller performance degradation percentage when it runs on a big number of nodes.} + Plots (\ref{fig:energy} and \ref{fig:per_deg}) present the energy saving and performance degradation respectively for all the benchmarks according to the number of used +nodes. As shown in the first plot, the energy saving percentages of the benchmarks MG, LU, BT and FT are decreased linearly when the the number of nodes is increased. While for the EP and SP benchmarks, the energy saving percentage is not affected by the increase of the number of computing nodes, because in these benchmarks there are no communications. Finally, the energy saving of the GC benchmark is significantly decreased when the number of nodes is increased because this benchmark has more communications than the others. The second plot shows that the performance degradation percentages of most of the benchmarks are decreased when they run on a big number of nodes because they spend more time communicating than computing, thus, scaling down the frequencies of some nodes have less effect on the performance. + + \subsection{The results for different power consumption scenarios} @@ -672,7 +674,7 @@ section, these ratios are changed and two new power scenarios are considered in \item 70\% dynamic power and 30\% static power \item 90\% dynamic power and 10\% static power \end{itemize} -The NAS parallel benchmarks were executed again over processors that follow the the new power scenarios. The class C of each benchmark was run over 8 or 9 nodes and the results are presented in tables (\ref{table:res_s1} and \ref{table:res_s2}). \textbf{These tables show that the energy saving percentage of the 70\%-30\% scenario is less for all benchmarks compared to the energy saving of the 90\%-10\% scenario, because this scenario uses higher percentage of dynamic dynamic power that is quadratically related to scaling factors. While the performance degradation percentage is less in 70\%-30\% scenario compared to 90\%-10\% scenario, because the first scenario used higher percentage for static power consumption that is linearly related to scaling factors and thus the execution time. } +The NAS parallel benchmarks were executed again over processors that follow the the new power scenarios. The class C of each benchmark was run over 8 or 9 nodes and the results are presented in tables (\ref{table:res_s1} and \ref{table:res_s2}). These tables show that the energy saving percentage of the 70\%-30\% scenario is less for all benchmarks compared to the energy saving of the 90\%-10\% scenario. Indeed, in the latter more dynamic power is consumed when nodes are running on their maximum frequencies, thus, scaling down the frequency of the nodes results in higher energy savings than in the 70\%-30\% scenario. On the other hand, the performance degradation percentage is less in the 70\%-30\% scenario compared to the 90\%-10\% scenario. This is due to the higher static power percentage in the first scenario which makes it more relevant in the overall consumed energy. Indeed, the static energy is related to the execution time and if the performance is degraded the total consumed static energy is directly increased. Therefore, the proposed algorithm do not scales down much the frequencies of the nodes in order to limit the increase of the execution time and thus limiting the effect of the consumed static energy . The two new power scenarios are compared to the old one in figure (\ref{fig:sen_comp}). It shows the average of the performance degradation, the energy saving and the distances for all NAS benchmarks of class C running on 8 or 9 nodes. The comparison shows that the energy saving ratio is proportional to the dynamic power ratio: it is increased when applying the 90\%-10\% scenario because at maximum frequency the dynamic energy is the the most relevant in the overall consumed energy and can be reduced by lowering the frequency of some processors. On the other hand, the energy saving is decreased when the 70\%-30\% scenario is used because the dynamic energy is less relevant in the overall consumed energy and lowering the frequency do not returns big energy savings. Moreover, the average of the performance degradation is decreased when using a higher ratio for static power (e.g. 70\%-30\% scenario and 80\%-20\% scenario). Since the proposed algorithm optimizes the energy consumption when using a higher ratio for dynamic power the algorithm selects bigger frequency scaling factors that result in more energy saving but less performance, for example see the figure (\ref{fig:scales_comp}). The opposite happens when using a higher ratio for static power, the algorithm proportionally selects smaller scaling values which results in less energy saving but less performance degradation.