1 \documentclass[conference]{IEEEtran}
3 \usepackage[T1]{fontenc}
4 \usepackage[utf8]{inputenc}
5 \usepackage[english]{babel}
6 \usepackage{algpseudocode}
12 \DeclareUrlCommand\email{\urlstyle{same}}
14 \usepackage[autolanguage,np]{numprint}
16 \renewcommand*\npunitcommand[1]{\text{#1}}
17 \npthousandthpartsep{}}
20 \usepackage[textsize=footnotesize]{todonotes}
21 \newcommand{\AG}[2][inline]{%
22 \todo[color=green!50,#1]{\sffamily\textbf{AG:} #2}\xspace}
23 \newcommand{\JC}[2][inline]{%
24 \todo[color=red!10,#1]{\sffamily\textbf{JC:} #2}\xspace}
26 \newcommand{\Xsub}[2]{{\ensuremath{#1_\mathit{#2}}}}
28 %% used to put some subscripts lower, and make them more legible
29 \newcommand{\fxheight}[1]{\ifx#1\relax\relax\else\rule{0pt}{1.52ex}#1\fi}
31 \newcommand{\CL}{\Xsub{C}{L}}
32 \newcommand{\Dist}{\mathit{Dist}}
33 \newcommand{\EdNew}{\Xsub{E}{dNew}}
34 \newcommand{\Eind}{\Xsub{E}{ind}}
35 \newcommand{\Enorm}{\Xsub{E}{Norm}}
36 \newcommand{\Eoriginal}{\Xsub{E}{Original}}
37 \newcommand{\Ereduced}{\Xsub{E}{Reduced}}
38 \newcommand{\Es}{\Xsub{E}{S}}
39 \newcommand{\Fdiff}[1][]{\Xsub{F}{diff}_{\!#1}}
40 \newcommand{\Fmax}[1][]{\Xsub{F}{max}_{\fxheight{#1}}}
41 \newcommand{\Fnew}{\Xsub{F}{new}}
42 \newcommand{\Ileak}{\Xsub{I}{leak}}
43 \newcommand{\Kdesign}{\Xsub{K}{design}}
44 \newcommand{\MaxDist}{\mathit{Max}\Dist}
45 \newcommand{\MinTcm}{\mathit{Min}\Tcm}
46 \newcommand{\Ntrans}{\Xsub{N}{trans}}
47 \newcommand{\Pd}[1][]{\Xsub{P}{d}_{\fxheight{#1}}}
48 \newcommand{\PdNew}{\Xsub{P}{dNew}}
49 \newcommand{\PdOld}{\Xsub{P}{dOld}}
50 \newcommand{\Pnorm}{\Xsub{P}{Norm}}
51 \newcommand{\Ps}[1][]{\Xsub{P}{s}_{\fxheight{#1}}}
52 \newcommand{\Scp}[1][]{\Xsub{S}{cp}_{#1}}
53 \newcommand{\Sopt}[1][]{\Xsub{S}{opt}_{#1}}
54 \newcommand{\Tcm}[1][]{\Xsub{T}{cm}_{\fxheight{#1}}}
55 \newcommand{\Tcp}[1][]{\Xsub{T}{cp}_{#1}}
56 \newcommand{\Ppeak}[1][]{\Xsub{P}{peak}_{#1}}
57 \newcommand{\Pidle}[1][]{\Xsub{P}{idle}_{\fxheight{#1}}}
58 \newcommand{\TcpOld}[1][]{\Xsub{T}{cpOld}_{#1}}
59 \newcommand{\Tnew}{\Xsub{T}{New}}
60 \newcommand{\Told}{\Xsub{T}{Old}}
64 \title{Energy Consumption Reduction with DVFS for \\
65 Message Passing Iterative Applications on \\
66 Heterogeneous Architectures}
76 FEMTO-ST Institute, University of Franche-Comté\\
77 IUT de Belfort-Montbéliard,
78 19 avenue du Maréchal Juin, BP 527, 90016 Belfort cedex, France\\
79 % Telephone: \mbox{+33 3 84 58 77 86}, % Raphaël
80 % Fax: \mbox{+33 3 84 58 77 81}\\ % Dept Info
81 Email: \email{{jean-claude.charr,raphael.couturier,ahmed.fanfakh_badri_muslim,arnaud.giersch}@univ-fcomte.fr}
92 \section{Introduction}
97 \section{Related works}
101 \section{The performance and energy consumption measurements on heterogeneous architecture}
104 \subsection{The execution time of message passing distributed iterative
105 applications on a heterogeneous platform}
107 In this paper, we are interested in reducing the energy consumption of message
108 passing distributed iterative synchronous applications running over
109 heterogeneous grid platforms. A heterogeneous grid platform could be defined as a collection of
110 heterogeneous computing clusters interconnected via a long distance network which has lower bandwidth
111 and higher latency than the local networks of the clusters. Each computing cluster in the grid is composed of homogeneous nodes that are connected together via high speed network. Therefore, each cluster has different characteristics such as computing power (FLOPS), energy consumption, CPU's frequency range, network bandwidth and latency.
115 \includegraphics[scale=0.6]{fig/commtasks}
116 \caption{Parallel tasks on a heterogeneous platform}
120 The overall execution time of a distributed iterative synchronous application
121 over a heterogeneous grid consists of the sum of the computation time and
122 the communication time for every iteration on a node. However, due to the
123 heterogeneous computation power of the computing clusters, slack times may occur
124 when fast nodes have to wait, during synchronous communications, for the slower
125 nodes to finish their computations (see Figure~\ref{fig:heter}). Therefore, the
126 overall execution time of the program is the execution time of the slowest task
127 which has the highest computation time and no slack time.
129 Dynamic Voltage and Frequency Scaling (DVFS) is a process, implemented in
130 modern processors, that reduces the energy consumption of a CPU by scaling
131 down its voltage and frequency. Since DVFS lowers the frequency of a CPU
132 and consequently its computing power, the execution time of a program running
133 over that scaled down processor may increase, especially if the program is
134 compute bound. The frequency reduction process can be expressed by the scaling
135 factor S which is the ratio between the maximum and the new frequency of a CPU
139 S = \frac{\Fmax}{\Fnew}
141 The execution time of a compute bound sequential program is linearly
142 proportional to the frequency scaling factor $S$. On the other hand, message
143 passing distributed applications consist of two parts: computation and
144 communication. The execution time of the computation part is linearly
145 proportional to the frequency scaling factor $S$ but the communication time is
146 not affected by the scaling factor because the processors involved remain idle
147 during the communications~\cite{Freeh_Exploring.the.Energy.Time.Tradeoff}. The
148 communication time for a task is the summation of periods of time that begin
149 with an MPI call for sending or receiving a message until the message is
150 synchronously sent or received.
152 Since in a heterogeneous grid each cluster has different characteristics,
153 especially different frequency gears, when applying DVFS operations on the nodes
154 of these clusters, they may get different scaling factors represented by a scaling vector:
155 $(S_{11}, S_{12},\dots, S_{NM})$ where $S_{ij}$ is the scaling factor of processor $j$ in cluster $i$ . To
156 be able to predict the execution time of message passing synchronous iterative
157 applications running over a heterogeneous grid, for different vectors of
158 scaling factors, the communication time and the computation time for all the
159 tasks must be measured during the first iteration before applying any DVFS
160 operation. Then the execution time for one iteration of the application with any
161 vector of scaling factors can be predicted using (\ref{eq:perf}).
164 \Tnew = \mathop{\max_{i=1,\dots N}}_{j=1,\dots,M}({\TcpOld[ij]} \cdot S_{ij})
165 +\mathop{\min_{j=1,\dots,M}} (\Tcm[hj])
168 where $N$ is the number of clusters in the grid, $M$ is the number of nodes in
169 each cluster, $\TcpOld[ij]$ is the computation time of processor $j$ in the cluster $i$
170 and $\Tcm[hj]$ is the communication time of processor $j$ in the cluster $h$ during the
171 first iteration. The model computes the maximum computation time with scaling factor
172 from each node added to the communication time of the slowest node in the slowest cluster $h$.
173 It means only the communication time without any slack time is taken into account.
174 Therefore, the execution time of the iterative application is equal to
175 the execution time of one iteration as in (\ref{eq:perf}) multiplied by the
176 number of iterations of that application.
178 This prediction model is developed from the model to predict the execution time
179 of message passing distributed applications for homogeneous and heterogeneous clusters
180 ~\cite{Our_first_paper,pdsec2015}. The execution time prediction model is
181 used in the method to optimize both the energy consumption and the performance
182 of iterative methods, which is presented in the following sections.
185 \subsection{Energy model for heterogeneous platform}
187 Many researchers~\cite{Malkowski_energy.efficient.high.performance.computing,
188 Rauber_Analytical.Modeling.for.Energy,Zhuo_Energy.efficient.Dynamic.Task.Scheduling,
189 Rizvandi_Some.Observations.on.Optimal.Frequency} divide the power consumed by
190 a processor into two power metrics: the static and the dynamic power. While the
191 first one is consumed as long as the computing unit is turned on, the latter is
192 only consumed during computation times. The dynamic power $\Pd$ is related to
193 the switching activity $\alpha$, load capacitance $\CL$, the supply voltage $V$
194 and operational frequency $F$, as shown in (\ref{eq:pd}).
197 \Pd = \alpha \cdot \CL \cdot V^2 \cdot F
199 The static power $\Ps$ captures the leakage power as follows:
202 \Ps = V \cdot \Ntrans \cdot \Kdesign \cdot \Ileak
204 where V is the supply voltage, $\Ntrans$ is the number of transistors,
205 $\Kdesign$ is a design dependent parameter and $\Ileak$ is a
206 technology dependent parameter. The energy consumed by an individual processor
207 to execute a given program can be computed as:
210 \Eind = \Pd \cdot \Tcp + \Ps \cdot T
212 where $T$ is the execution time of the program, $\Tcp$ is the computation
213 time and $\Tcp \le T$. $\Tcp$ may be equal to $T$ if there is no
214 communication and no slack time.
216 The main objective of DVFS operation is to reduce the overall energy
217 consumption~\cite{Le_DVFS.Laws.of.Diminishing.Returns}. The operational
218 frequency $F$ depends linearly on the supply voltage $V$, i.e., $V = \beta \cdot
219 F$ with some constant $\beta$.~This equation is used to study the change of the
220 dynamic voltage with respect to various frequency values
221 in~\cite{Rauber_Analytical.Modeling.for.Energy}. The reduction process of the
222 frequency can be expressed by the scaling factor $S$ which is the ratio between
223 the maximum and the new frequency as in (\ref{eq:s}). The CPU governors are
224 power schemes supplied by the operating system's kernel to lower a core's
225 frequency. The new frequency $\Fnew$ from (\ref{eq:s}) can be calculated as
229 \Fnew = S^{-1} \cdot \Fmax
231 Replacing $\Fnew$ in (\ref{eq:pd}) as in (\ref{eq:fnew}) gives the following
232 equation for dynamic power consumption:
235 \PdNew = \alpha \cdot \CL \cdot V^2 \cdot \Fnew = \alpha \cdot \CL \cdot \beta^2 \cdot \Fnew^3 \\
236 {} = \alpha \cdot \CL \cdot V^2 \cdot \Fmax \cdot S^{-3} = \PdOld \cdot S^{-3}
238 where $\PdNew$ and $\PdOld$ are the dynamic power consumed with the
239 new frequency and the maximum frequency respectively.
241 According to (\ref{eq:pdnew}) the dynamic power is reduced by a factor of
242 $S^{-3}$ when reducing the frequency by a factor of
243 $S$~\cite{Rauber_Analytical.Modeling.for.Energy}. Since the FLOPS of a CPU is
244 proportional to the frequency of a CPU, the computation time is increased
245 proportionally to $S$. The new dynamic energy is the dynamic power multiplied
246 by the new time of computation and is given by the following equation:
249 \EdNew = \PdOld \cdot S^{-3} \cdot (\Tcp \cdot S)= S^{-2}\cdot \PdOld \cdot \Tcp
251 The static power is related to the power leakage of the CPU and is consumed
252 during computation and even when idle. As
253 in~\cite{Rauber_Analytical.Modeling.for.Energy,Zhuo_Energy.efficient.Dynamic.Task.Scheduling},
254 the static power of a processor is considered as constant during idle and
255 computation periods, and for all its available frequencies. The static energy
256 is the static power multiplied by the execution time of the program. According
257 to the execution time model in (\ref{eq:perf}), the execution time of the
258 program is the sum of the computation and the communication times. The
259 computation time is linearly related to the frequency scaling factor, while this
260 scaling factor does not affect the communication time. The static energy of a
261 processor after scaling its frequency is computed as follows:
264 \Es = \Ps \cdot (\Tcp \cdot S + \Tcm)
267 In the considered heterogeneous grid platform, each node $j$ in cluster $i$ may have
268 different dynamic and static powers from the nodes of the other clusters,
269 noted as $\Pd[ij]$ and $\Ps[ij]$ respectively. Therefore, even if the distributed
270 message passing iterative application is load balanced, the computation time of each CPU $j$
271 in cluster $i$ noted $\Tcp[ij]$ may be different and different frequency scaling factors may be
272 computed in order to decrease the overall energy consumption of the application
273 and reduce slack times. The communication time of a processor $j$ in cluster $i$ is noted as
274 $\Tcm[ij]$ and could contain slack times when communicating with slower nodes,
275 see Figure~\ref{fig:heter}. Therefore, all nodes do not have equal
276 communication times. While the dynamic energy is computed according to the
277 frequency scaling factor and the dynamic power of each node as in
278 (\ref{eq:Edyn}), the static energy is computed as the sum of the execution time
279 of one iteration multiplied by the static power of each processor. The overall
280 energy consumption of a message passing distributed application executed over a
281 heterogeneous grid platform during one iteration is the summation of all dynamic and
282 static energies for $M$ processors in $N$ clusters. It is computed as follows:
285 E = \sum_{i=1}^{N} \sum_{i=1}^{M} {(S_{ij}^{-2} \cdot \Pd[ij] \cdot \Tcp[ij])} +
286 \sum_{i=1}^{N} \sum_{j=1}^{M} (\Ps[ij] \cdot {} \\
287 (\mathop{\max_{i=1,\dots N}}_{j=1,\dots,M}({\Tcp[ij]} \cdot S_{ij})
288 +\mathop{\min_{j=1,\dots M}} (\Tcm[hj]) ))
291 Reducing the frequencies of the processors according to the vector of scaling
292 factors $(S_{11}, S_{12},\dots, S_{NM})$ may degrade the performance of the application
293 and thus, increase the static energy because the execution time is
294 increased~\cite{Kim_Leakage.Current.Moore.Law}. The overall energy consumption
295 for the iterative application can be measured by measuring the energy
296 consumption for one iteration as in (\ref{eq:energy}) multiplied by the number
297 of iterations of that application.
299 \section{Optimization of both energy consumption and performance}
302 Using the lowest frequency for each processor does not necessarily give the most
303 energy efficient execution of an application. Indeed, even though the dynamic
304 power is reduced while scaling down the frequency of a processor, its
305 computation power is proportionally decreased. Hence, the execution time might
306 be drastically increased and during that time, dynamic and static powers are
307 being consumed. Therefore, it might cancel any gains achieved by scaling down
308 the frequency of all nodes to the minimum and the overall energy consumption of
309 the application might not be the optimal one. It is not trivial to select the
310 appropriate frequency scaling factor for each processor while considering the
311 characteristics of each processor (computation power, range of frequencies,
312 dynamic and static powers) and the task executed (computation/communication
313 ratio). The aim being to reduce the overall energy consumption and to avoid
314 increasing significantly the execution time. In our previous
315 work~\cite{Our_first_paper,pdsec2015}, we proposed a method that selects the optimal
316 frequency scaling factor for a homogeneous and heterogeneous clusters executing a message passing
317 iterative synchronous application while giving the best trade-off between the
318 energy consumption and the performance for such applications. In this work we
319 are interested in heterogeneous grid as described above. Due to the
320 heterogeneity of the processors, a vector of scaling factors should be selected
321 and it must give the best trade-off between energy consumption and performance.
323 The relation between the energy consumption and the execution time for an
324 application is complex and nonlinear, Thus, unlike the relation between the
325 execution time and the scaling factor, the relation between the energy and the
326 frequency scaling factors is nonlinear, for more details refer
327 to~\cite{Freeh_Exploring.the.Energy.Time.Tradeoff}. Moreover, these relations
328 are not measured using the same metric. To solve this problem, the execution
329 time is normalized by computing the ratio between the new execution time (after
330 scaling down the frequencies of some processors) and the initial one (with
331 maximum frequency for all nodes) as follows:
334 \Pnorm = \frac{\Tnew}{\Told}
338 Where $Tnew$ is computed as in (\ref{eq:perf}) and $Told$ is computed as in (\ref{eq:told})
341 \Told = \mathop{\max_{i=1,2,\dots,N}}_{j=1,2,\dots,M} (\Tcp[ij]+\Tcm[ij])
343 In the same way, the energy is normalized by computing the ratio between the
344 consumed energy while scaling down the frequency and the consumed energy with
345 maximum frequency for all nodes:
348 \Enorm = \frac{\Ereduced}{\Eoriginal}
351 Where $\Ereduced$ is computed using (\ref{eq:energy}) and $\Eoriginal$ is
354 \textcolor{red}{A reference is missing}
357 \Eoriginal = \sum_{i=1}^{N} \sum_{j=1}^{M} ( \Pd[ij] \cdot \Tcp[ij]) +
358 \mathop{\sum_{i=1}^{N}} \sum_{j=1}^{M} (\Ps[ij] \cdot \Told)
361 While the main goal is to optimize the energy and execution time at the same
362 time, the normalized energy and execution time curves do not evolve (increase/decrease) in the same way.
363 According to the equations~(\ref{eq:pnorm}) and (\ref{eq:enorm}), the
364 vector of frequency scaling factors $S_1,S_2,\dots,S_N$ reduce both the energy
365 and the execution time simultaneously. But the main objective is to produce
366 maximum energy reduction with minimum execution time reduction.
368 This problem can be solved by making the optimization process for energy and
369 execution time follow the same evolution according to the vector of scaling factors
370 $(S_{11}, S_{12},\dots, S_{NM})$. Therefore, the equation of the
371 normalized execution time is inverted which gives the normalized performance
372 equation, as follows:
375 \Pnorm = \frac{\Told}{\Tnew}
380 \subfloat[Homogeneous cluster]{%
381 \includegraphics[width=.33\textwidth]{fig/homo}\label{fig:r1}}%
383 \subfloat[Heterogeneous grid]{%
384 \includegraphics[width=.33\textwidth]{fig/heter}\label{fig:r2}}
386 \caption{The energy and performance relation}
389 Then, the objective function can be modeled in order to find the maximum
390 distance between the energy curve (\ref{eq:enorm}) and the performance curve
391 (\ref{eq:pnorm_inv}) over all available sets of scaling factors. This
392 represents the minimum energy consumption with minimum execution time (maximum
393 performance) at the same time, see Figure~\ref{fig:r1} or
394 Figure~\ref{fig:r2}. Then the objective function has the following form:
398 \mathop{ \mathop{\max_{i=1,\dots N}}_{j=1,\dots,M}}_{k=1,\dots,F}
399 (\overbrace{\Pnorm(S_{ijk})}^{\text{Maximize}} -
400 \overbrace{\Enorm(S_{ijk})}^{\text{Minimize}} )
402 where $N$ is the number of clusters, $M$ is the number of nodes in each cluster and
403 $F$ is the number of available frequencies for each node. Then, the optimal set
404 of scaling factors that satisfies (\ref{eq:max}) can be selected.
405 The objective function can work with any energy model or any power
406 values for each node (static and dynamic powers). However, the most important
407 energy reduction gain can be achieved when the energy curve has a convex form as shown
408 in~\cite{Zhuo_Energy.efficient.Dynamic.Task.Scheduling,Rauber_Analytical.Modeling.for.Energy,Hao_Learning.based.DVFS}.
410 \section{The scaling factors selection algorithm for grids }
414 \begin{algorithmic}[1]
418 \item [{$N$}] number of clusters in the grid.
419 \item [{$M$}] number of nodes in each cluster.
420 \item[{$\Tcp[ij]$}] array of all computation times for all nodes during one iteration and with the highest frequency.
421 \item[{$\Tcm[ij]$}] array of all communication times for all nodes during one iteration and with the highest frequency.
422 \item[{$\Fmax[ij]$}] array of the maximum frequencies for all nodes.
423 \item[{$\Pd[ij]$}] array of the dynamic powers for all nodes.
424 \item[{$\Ps[ij]$}] array of the static powers for all nodes.
425 \item[{$\Fdiff[ij]$}] array of the differences between two successive frequencies for all nodes.
427 \Ensure $\Sopt[11],\Sopt[12] \dots, \Sopt[NM_i]$, a vector of scaling factors that gives the optimal tradeoff between energy consumption and execution time
429 \State $\Scp[ij] \gets \frac{\max_{i=1,2,\dots,N}(\max_{j=1,2,\dots,M_i}(\Tcp[ij]))}{\Tcp[ij]} $
430 \State $F_{ij} \gets \frac{\Fmax[ij]}{\Scp[i]},~{i=1,2,\cdots,N},~{j=1,2,\dots,M_i}.$
431 \State Round the computed initial frequencies $F_i$ to the closest available frequency for each node.
432 \If{(not the first frequency)}
433 \State $F_{ij} \gets F_{ij}+\Fdiff[ij],~i=1,\dots,N,~{j=1,\dots,M_i}.$
435 \State $\Told \gets $ computed as in equations (\ref{eq:told}).
436 \State $\Eoriginal \gets $ computed as in equations (\ref{eq:eorginal}) .
437 \State $\Sopt[ij] \gets 1,~i=1,\dots,N,~{j=1,\dots,M_i}. $
438 \State $\Dist \gets 0 $
439 \While {(all nodes have not reached their minimum \newline\hspace*{2.5em} frequency \textbf{or} $\Pnorm - \Enorm < 0 $)}
440 \If{(not the last freq. \textbf{and} not the slowest node)}
441 \State $F_{ij} \gets F_{ij} - \Fdiff[ij],~{i=1,\dots,N},~{j=1,\dots,M_i}$.
442 \State $S_{ij} \gets \frac{\Fmax[ij]}{F_{ij}},~{i=1,\dots,N},~{j=1,\dots,M_i}.$
444 \State $\Tnew \gets $ computed as in equations (\ref{eq:perf}).
445 \State $\Ereduced \gets $ computed as in equations (\ref{eq:energy}).
446 \State $\Pnorm \gets \frac{\Told}{\Tnew}$
447 \State $\Enorm\gets \frac{\Ereduced}{\Eoriginal}$
448 \If{$(\Pnorm - \Enorm > \Dist)$}
449 \State $\Sopt[ij] \gets S_{ij},~i=1,\dots,N,~j=1,\dots,M_i. $
450 \State $\Dist \gets \Pnorm - \Enorm$
453 \State Return $\Sopt[11],\Sopt[12],\dots,\Sopt[NM_i]$
455 \caption{Scaling factors selection algorithm}
460 \begin{algorithmic}[1]
462 \For {$k=1$ to \textit{some iterations}}
463 \State Computations section.
464 \State Communications section.
466 \State Gather all times of computation and\newline\hspace*{3em}%
467 communication from each node.
468 \State Call Algorithm \ref{HSA}.
469 \State Compute the new frequencies from the\newline\hspace*{3em}%
470 returned optimal scaling factors.
471 \State Set the new frequencies to nodes.
475 \caption{DVFS algorithm}
479 \subsection{The algorithm details}
481 \textcolor{red}{Delete the subsection if there's only one.}
483 In this section, the scaling factors selection algorithm for grids, algorithm~\ref{HSA}, is presented. It selects the vector of the frequency
484 scaling factors that gives the best trade-off between minimizing the
485 energy consumption and maximizing the performance of a message passing
486 synchronous iterative application executed on a grid. It works
487 online during the execution time of the iterative message passing program. It
488 uses information gathered during the first iteration such as the computation
489 time and the communication time in one iteration for each node. The algorithm is
490 executed after the first iteration and returns a vector of optimal frequency
491 scaling factors that satisfies the objective function (\ref{eq:max}). The
492 program applies DVFS operations to change the frequencies of the CPUs according
493 to the computed scaling factors. This algorithm is called just once during the
494 execution of the program. Algorithm~\ref{dvfs} shows where and when the proposed
495 scaling algorithm is called in the iterative MPI program.
499 \includegraphics[scale=0.45]{fig/init_freq}
500 \caption{Selecting the initial frequencies}
504 Nodes from distinct clusters in a grid have different computing powers, thus
505 while executing message passing iterative synchronous applications, fast nodes
506 have to wait for the slower ones to finish their computations before being able
507 to synchronously communicate with them as in Figure~\ref{fig:heter}. These
508 periods are called idle or slack times. The algorithm takes into account this
509 problem and tries to reduce these slack times when selecting the vector of the frequency
510 scaling factors. At first, it selects initial frequency scaling factors
511 that increase the execution times of fast nodes and minimize the differences
512 between the computation times of fast and slow nodes. The value of the initial
513 frequency scaling factor for each node is inversely proportional to its
514 computation time that was gathered from the first iteration. These initial
515 frequency scaling factors are computed as a ratio between the computation time
516 of the slowest node and the computation time of the node $i$ as follows:
519 \Scp[ij] = \frac{ \mathop{\max_{i=1,\dots N}}_{j=1,\dots,M}(\Tcp[ij])} {\Tcp[ij]}
521 Using the initial frequency scaling factors computed in (\ref{eq:Scp}), the
522 algorithm computes the initial frequencies for all nodes as a ratio between the
523 maximum frequency of node $i$ and the computation scaling factor $\Scp[i]$ as
527 F_{ij} = \frac{\Fmax[ij]}{\Scp[ij]},~{i=1,2,\dots,N},~{j=1,\dots,M}
529 If the computed initial frequency for a node is not available in the gears of
530 that node, it is replaced by the nearest available frequency. In
531 Figure~\ref{fig:st_freq}, the nodes are sorted by their computing powers in
532 ascending order and the frequencies of the faster nodes are scaled down
533 according to the computed initial frequency scaling factors. The resulting new
534 frequencies are highlighted in Figure~\ref{fig:st_freq}. This set of
535 frequencies can be considered as a higher bound for the search space of the
536 optimal vector of frequencies because selecting higher frequencies
537 than the higher bound will not improve the performance of the application and it
538 will increase its overall energy consumption. Therefore the algorithm that
539 selects the frequency scaling factors starts the search method from these
540 initial frequencies and takes a downward search direction toward lower
541 frequencies until reaching the nodes' minimum frequencies or lower bounds. A node's frequency is considered its lower bound if the computed distance between the energy and performance at this frequency is less than zero.
542 A negative distance means that the performance degradation ratio is higher than the energy saving ratio.
543 In this situation, the algorithm must stop the downward search because it has reached the lower bound and it is useless to test the lower frequencies. Indeed, they will all give worse distances.
545 Therefore, the algorithm iterates on all remaining frequencies, from the higher
546 bound until all nodes reach their minimum frequencies or their lower bounds, to compute the overall
547 energy consumption and performance and selects the optimal vector of the frequency scaling
548 factors. At each iteration the algorithm determines the slowest node
549 according to the equation (\ref{eq:perf}) and keeps its frequency unchanged,
550 while it lowers the frequency of all other nodes by one gear. The new overall
551 energy consumption and execution time are computed according to the new scaling
552 factors. The optimal set of frequency scaling factors is the set that gives the
553 highest distance according to the objective function (\ref{eq:max}).
555 Figures~\ref{fig:r1} and \ref{fig:r2} illustrate the normalized performance and
556 consumed energy for an application running on a homogeneous cluster and a
557 grid platform respectively while increasing the scaling factors. It can
558 be noticed that in a homogeneous cluster the search for the optimal scaling
559 factor should start from the maximum frequency because the performance and the
560 consumed energy decrease from the beginning of the plot. On the other hand, in
561 the grid platform the performance is maintained at the beginning of the
562 plot even if the frequencies of the faster nodes decrease until the computing
563 power of scaled down nodes are lower than the slowest node. In other words,
564 until they reach the higher bound. It can also be noticed that the higher the
565 difference between the faster nodes and the slower nodes is, the bigger the
566 maximum distance between the energy curve and the performance curve is, which results in bigger energy savings.
569 \section{Experimental results}
571 While in~\cite{mpi-energy2} the energy model and the scaling factors selection algorithm were applied to a heterogeneous cluster and evaluated over the SimGrid simulator~\cite{SimGrid.org},
572 in this paper real experiments were conducted over the grid'5000 platform.
574 \subsection{Grid'5000 architature and power consumption}
576 Grid'5000~\cite{grid5000} is a large-scale testbed that consists of ten sites distributed over all metropolitan France and Luxembourg. All the sites are connected together via a special long distance network called RENATER,
577 which is the French National Telecommunication Network for Technology.
578 Each site of the grid is composed of few heterogeneous
579 computing clusters and each cluster contains many homogeneous nodes. In total,
580 grid'5000 has about one thousand heterogeneous nodes and eight thousand cores. In each site,
581 the clusters and their nodes are connected via high speed local area networks.
582 Two types of local networks are used, Ethernet or Infiniband networks which have different characteristics in terms of bandwidth and latency.
584 Since grid'5000 is dedicated for testing, contrary to production grids it allows a user to deploy its own customized operating system on all the booked nodes. The user could have root rights and thus apply DVFS operations while executing a distributed application. Moreover, the grid'5000 testbed provides at some sites a power measurement tool to capture
585 the power consumption for each node in those sites. The measured power is the overall consumed power by by all the components of a node at a given instant, such as CPU, hard drive, main-board, memory, ... For more details refer to
586 \cite{Energy_measurement}. To just measure the CPU power of one core in a node $j$,
587 firstly, the power consumed by the node while being idle at instant $y$, noted as $\Pidle[jy]$, was measured. Then, the power was measured while running a single thread benchmark with no communication (no idle time) over the same node with its CPU scaled to the maximum available frequency. The latter power measured at time $x$ with maximum frequency for one core of node $j$ is noted $P\max[jx]$. The difference between the two measured power consumption represents the
588 dynamic power consumption of that core with the maximum frequency, see figure(\ref{fig:power_cons}).
590 \textcolor{red}{why maximum and minimum, change peak in the equation and the figure}
592 The dynamic power $\Pd[j]$ is computed as in equation (\ref{eq:pdyn})
595 \Pd[j] = \max_{x=\beta_1,\dots \beta_2} (P\max[jx]) - \min_{y=\Theta_1,\dots \Theta_2} (\Pidle[jy])
598 where $\Pd[j]$ is the dynamic power consumption for one core of node $j$,
599 $\lbrace \beta_1,\beta_2 \rbrace$ is the time interval for the measured peak power values,
600 $\lbrace\Theta_1,\Theta_2\rbrace$ is the time interval for the measured idle power values.
601 Therefore, the dynamic power of one core is computed as the difference between the maximum
602 measured value in peak powers vector and the minimum measured value in the idle powers vector.
604 On the other hand, the static power consumption by one core is a part of the measured idle power consumption of the node. Since in grid'5000 there is no way to measure precisely the consumed static power and in~\cite{Our_first_paper,pdsec2015,Rauber_Analytical.Modeling.for.Energy} it was assumed that the static power represents a ratio of the dynamic power, the value of the static power is assumed as np[\%]{20} of dynamic power consumption of the core.
606 In the experiments presented in the following sections, two sites of grid'5000 were used, Lyon and Nancy sites. These two sites have in total seven different clusters as in figure (\ref{fig:grid5000}).
608 Four clusters from the two sites were selected in the experiments: one cluster from
609 Lyon's site, Taurus cluster, and three clusters from Nancy's site, Graphene,
610 Griffon and Graphite. Each one of these clusters has homogeneous nodes inside, while nodes from different clusters are heterogeneous in many aspects such as: computing power, power consumption, available
611 frequency ranges and local network features: the bandwidth and the latency. Table \ref{table:grid5000} shows
612 the details characteristics of these four clusters. Moreover, the dynamic powers were computed using the equation (\ref{eq:pdyn}) for all the nodes in the
613 selected clusters and are presented in table \ref{table:grid5000}.
620 \includegraphics[scale=1]{fig/grid5000}
621 \caption{The selected two sites of grid'5000}
626 The energy model and the scaling factors selection algorithm were applied to the NAS parallel benchmarks v3.3 \cite{NAS.Parallel.Benchmarks} and evaluated over grid'5000.
627 The benchmark suite contains seven applications: CG, MG, EP, LU, BT, SP and FT. These applications have different computations and communications ratios and strategies which make them good testbed applications to evaluate the proposed algorithm and energy model.
628 The benchmarks have seven different classes, S, W, A, B, C, D and E, that represent the size of the problem that the method solves. In this work, the class D was used for all benchmarks in all the experiments presented in the next sections.
635 \includegraphics[scale=0.6]{fig/power_consumption.pdf}
636 \caption{The power consumption by one core from Taurus cluster}
637 \label{fig:power_cons}
644 \caption{CPUs characteristics of the selected clusters}
647 \begin{tabular}{|*{7}{c|}}
649 Cluster & CPU & Max & Min & Diff. & no. of cores & dynamic power \\
650 Name & model & Freq. & Freq. & Freq. & per CPU & of one core \\
651 & & GHz & GHz & GHz & & \\
653 Taurus & Intel & 2.3 & 1.2 & 0.1 & 6 & \np[W]{35} \\
655 & E5-2630 & & & & & \\
657 Graphene & Intel & 2.53 & 1.2 & 0.133 & 4 & \np[W]{23} \\
661 Griffon & Intel & 2.5 & 2 & 0.5 & 4 & \np[W]{46} \\
665 Graphite & Intel & 2 & 1.2 & 0.1 & 8 & \np[W]{35} \\
667 & E5-2650 & & & & & \\
670 \label{table:grid5000}
676 \subsection{The experimental results of the scaling algorithm}
679 \subsection{The experimental results of multi-cores clusters}
682 \subsection{The results for different power consumption scenarios}
688 \subsection{The comparison of the proposed scaling algorithm }
689 \label{sec.compare_EDP}
698 \section*{Acknowledgment}
700 This work has been partially supported by the Labex ACTION project (contract
701 ``ANR-11-LABX-01-01''). Computations have been performed on the supercomputer
702 facilities of the Mésocentre de calcul de Franche-Comté. As a PhD student,
703 Mr. Ahmed Fanfakh, would like to thank the University of Babylon (Iraq) for
706 % trigger a \newpage just before the given reference
707 % number - used to balance the columns on the last page
708 % adjust value as needed - may need to be readjusted if
709 % the document is modified later
710 %\IEEEtriggeratref{15}
712 \bibliographystyle{IEEEtran}
713 \bibliography{IEEEabrv,my_reference}
720 %%% ispell-local-dictionary: "american"
723 % LocalWords: Fanfakh Charr FIXME Tianhe DVFS HPC NAS NPB SMPI Rauber's Rauber
724 % LocalWords: CMOS EPSA Franche Comté Tflop Rünger IUT Maréchal Juin cedex GPU
725 % LocalWords: de badri muslim MPI SimGrid GFlops Xeon EP BT GPUs CPUs AMD
726 % LocalWords: Spiliopoulos scalability