\title{Answers to the questions of the reviewers}
\maketitle
-\section{Questions of the first reviewer}
+\section{Questions and remarks of the first reviewer}
\begin{enumerate}
savings, but your arguments for providing a solution for this was
based that today's supercomputers are achieving massive scale.
-\textbf{Answer:} In the Figure 7, the energy consumption of the benchmarks solving the class D and running on many scenarios are presented. The number of used nodes varies between 16 and 32 in the scenarios while the size of the problem is not modified. Therefore, the computations to communications times ratio is lower when 32 nodes are used instead of 16. When this ratio is small, it means there are not enough computations when compared to the communications times and the impact of scaling down the frequency of the CPU on its energy consumption is reduced. To solve this problem, the problem
+\textbf{Answer:} In Figure 7, the energy consumption of the benchmarks solving the class D and running on many scenarios are presented. The number of used nodes varies between 16 and 32 in the scenarios while the size of the problem is not modified. Therefore, the computations to communications times ratio is lower when 32 nodes are used instead of 16. When this ratio is small, it means there are not enough computations when compared to the communications times and the impact of scaling down the frequency of the CPU on its energy consumption is reduced. To solve this problem, the problem
should be solved on a number of nodes adequate to its size. For example, for the NAS benchmarks, the class E should have been solved on 32 nodes to have a good computations to communications times ratio.
\item In Sec 6.3, why did you choose to keep 32 processes for the evaluation
with multi-core clusters? How did you configure MPI for the results
-\textcolor{blue}{Answer: We keep choosing 32 nodes in both scenarios
-to compare them while one core per node scenario has distributed communications (one network link for each node) and multi-core scenario uses shared network link communications and thus comparing their impact on the results.
-We configure MPI on one core per node scenario by choosing one core per nodes (e.g in machine file we did: node1, node2 ,node3, node4). While in multi-core scenario we choose one machine with four cores (e.g. node1 slots=4).}
+\textbf{Answer:} In section 6.3, we wanted to evaluate how much energy can be saved when applying the proposed algorithm to message passing applications with iterations running over a grid composed of multi-core nodes. Therefore, the same experiments as in section 6.2 were conducted on the new multi-core platform. Instead of running one process per node as in the previous section, 3 or 4 processes were executed on each multi-core node. The total number of processes, 32 processes, was not modified in order to fairly compare the single core and the multi-core versions.
+
+Only the architecture file was modified between the single and the multi-core architectures. For the single core architecture, the architecture file contains the name of 32 different nodes. For the multi-core architecture, the architecture file contains less nodes and for every node 3 or 4 slots (cores) are used. The total number of slots is equal to 32.
+
\item shown in Fig 8a? Some MPI implementations have an option to use shared
shared-memory optimization is possible (sometimes automatically
detected by MPI if you pin processes to cores).
-\textcolor{blue}{Answer: We didn't manually pin processes to cores and since the communication times
-increased. We guess that the shared memory wasn't used.}
+\textbf{Answer:} We did not manually pin processes to cores. Since the communication times
+increased, we think that the shared memory was not used when two processes, running on the same node, exchange data.
\item In P33, Sec 6.5, you mention that the proposed algorithm outperforms
EDP because the former considers both metrics (time, energy) and the
defined: energy x execution time. I think this is only a matter of
phrasing.
-\textcolor{blue}{Answer: we use the delay in execution time not the execution time. Then, the equation that we used is EDP= energy x (Tnew-Told). The experiments shows that our objective function
-is better than the EDP objective.}
+ \textbf{Answer:} We agree with the reviewer, EDP also uses two metrics in the objective function: energy and delay. The sentence in the paper was modified to clarify this misunderstanding. The main difference between our algorithm and the EDP method is the used objective function. For EDP, the product of energy and delay must be minimized, while for our algorithm, the difference between the normalized performance and the normalized energy should be maximized. This new formulation of the objective function allows our algorithm to select the set of frequencies that gives the best tradeoff between the energy consumption and the performance. The objective function of EDP does not give the same frequencies as our algorithm and thus it is outperformed by our method. The results of the experiments confirm that the objective function used by our algorithm is more efficient than the one used by EDP.
\item Other complementary points to consider:
+ Same for Fig 7.
-\item \textcolor{blue}{ Answer: We have considered these points in the paper.}
+ \textbf{Answer:} Answer: We have taken in consideration all these remarks and the paper was modified accordingly.
\item From the design of experiments, did you consider using replications?
There is no variability metric in your results. Have you run multiple
feel that such variability needs to be accounted for, otherwise it is
very hard to affirm anything about measurements.
- \textcolor{blue}{Answer: Each experiment has been executed many times and the results presented in the
- figures are the average values of many executions.}
+ \textbf{Answer:} Each experiment has been executed many times and the results presented in the
+ figures are the average values of many executions. Since we have deployed the same operating system on the booked machines and we were the only users executing processes on them during the experiments, no significant variability in the execution time of the applications was noticed.
\item In summary, I think this is a very interesting work but the experimental evaluation lacks variability measurements, consider larger experiments (1K nodes for instance) to see how everything scales, and there is no overhead measurements although authors stress that in abstract/introduction.
-\textcolor{blue}{Answer: We will expand the experimental over a large number of nodes in the future work while increasing the problem size and considering the variability issues. We have discussed the algorithm overhead and its complexity in section 6.5.}
-\end{enumerate}
+\textbf{Answer:} For the time being, we do not have the resources nor the time to evaluate the proposed algorithm over large platforms composed of more than 1K nodes. However, as said in the perspectives of the paper, the evaluation of the scalability of the algorithm will be in a conducted in a future work as soon as we have access to larger resources. We have discussed the overhead of the algorithm and its complexity in section 6.5 and given in the answer to question 2 some solutions to improve its scalability and reduce its overhead.
+For the variability issue, please refer to the answer to question 1.
-\section{Questions of the second reviewer}
-This paper presents detailed performance and energy model for iterative message passing applications. Further a method is proposed to select the frequencies of heterogeneous cpus online. The selection method itself is not difficult. But I like the systematic modeling for energy consumption and performance. This paper is well written in general. The technical contents are presented in a logical way overall. The experiments are conducted in real platform, which shows the practicality of this work and also makes the work have more impact on the field. However, I have the following comments and concerns for this paper. The authors should clarify them in the revised version.
+
+\end{enumerate}
--move the contributions from related work to introduction
+\section{Questions and remarks of the second reviewer}
\begin{enumerate}
-
+\item Move the contributions from related work to introduction
\item why emphasize it is a grid platform? the presentation of related work follows the logic of heterogeneous CPUs. Grid is only a type of platform with heterogeneous CPUs
The proposed method can be applied to a heterogeneous platform composed from GPUs and CPUs, since modern GPUs like CPUs allow the use of DVFS operation.}
\end{enumerate}
-\section{Questions of the third reviewer}
-In this paper, a new online frequency selecting algorithm for grids, composed of heterogeneous clusters, is presented. It selects the frequencies and tries to give the best trade-off between
-energy saving and performance degradation, for each node computing the message passing iterative application. The algorithm has a small overhead and works without training or profiling. It uses a new energy model for message
-passing iterative applications running on a grid. The proposed algorithm is evaluated on a real grid, the Grid'5000 platform, while running the NAS parallel benchmarks. The experiments on 16 nodes, distributed on three clusters, show
-that it reduces on average the energy consumption by 30\% while the performance is on average only degraded by 3.2\%. Finally, the algorithm is compared to an existing method. The comparison results show that it outperforms the
-latter in terms of energy consumption reduction and performance.
-
-this paper is quite interesting and solid. But before acceptance, I suggest to have the following major revisions:
+\section{Questions and remarks of the third reviewer}
\begin{enumerate}
\item suggest the authors to use much larger size of nodes, instead of on 16 nodes, distributed on three clusters, to see the scalability of the energy saving