1 \documentclass[smallextended]{svjour3}
2 \usepackage[utf8]{inputenc}
3 \usepackage[T1]{fontenc}
8 \usepackage[ruled,lined]{algorithm2e}
10 \newcommand{\abs}[1]{\lvert#1\rvert} % \abs{x} -> |x|
12 \newenvironment{algodata}{%
13 \begin{tabular}[t]{@{}l@{:~}l@{}}}{%
16 \newcommand{\VAR}[1]{\textit{#1}}
20 \title{Best effort strategy and virtual load
21 for asynchronous iterative load balancing}
23 \author{Raphaël Couturier \and
27 \institute{R. Couturier \and A. Giersch \at
28 LIFC, University of Franche-Comté, Belfort, France \\
29 % Tel.: +123-45-678910\\
30 % Fax: +123-45-678910\\
32 raphael.couturier@univ-fcomte.fr,
33 arnaud.giersch@univ-fcomte.fr}
41 Most of the time, asynchronous load balancing algorithms have extensively been
42 studied in a theoretical point of view. The Bertsekas and Tsitsiklis'
43 algorithm~\cite[section~7.4]{bertsekas+tsitsiklis.1997.parallel}
44 is certainly the most well known algorithm for which the convergence proof is
45 given. From a practical point of view, when a node wants to balance a part of
46 its load to some of its neighbors, the strategy is not described. In this
47 paper, we propose a strategy called \emph{best effort} which tries to balance
48 the load of a node to all its less loaded neighbors while ensuring that all the
49 nodes concerned by the load balancing phase have the same amount of load.
50 Moreover, asynchronous iterative algorithms in which an asynchronous load
51 balancing algorithm is implemented most of the time can dissociate messages
52 concerning load transfers and message concerning load information. In order to
53 increase the converge of a load balancing algorithm, we propose a simple
54 heuristic called \emph{virtual load} which allows a node that receives a load
55 information message to integrate the load that it will receive later in its
56 load (virtually) and consequently sends a (real) part of its load to some of its
57 neighbors. In order to validate our approaches, we have defined a simulator
58 based on SimGrid which allowed us to conduct many experiments.
63 \section{Introduction}
65 Load balancing algorithms are extensively used in parallel and distributed
66 applications in order to reduce the execution times. They can be applied in
67 different scientific fields from high performance computation to micro sensor
68 networks. They are iterative by nature. In literature many kinds of load
69 balancing algorithms have been studied. They can be classified according
70 different criteria: centralized or decentralized, in static or dynamic
71 environment, with homogeneous or heterogeneous load, using synchronous or
72 asynchronous iterations, with a static topology or a dynamic one which evolves
73 during time. In this work, we focus on asynchronous load balancing algorithms
74 where computer nodes are considered homogeneous and with homogeneous load with
75 no external load. In this context, Bertsekas and Tsitsiklis have proposed an
76 algorithm which is definitively a reference for many works. In their work, they
77 proved that under classical hypotheses of asynchronous iterative algorithms and
78 a special constraint avoiding \emph{ping-pong} effect, an asynchronous
79 iterative algorithm converge to the uniform load distribution. This work has
80 been extended by many authors. For example, Cortés et al., with
81 DASUD~\cite{cortes+ripoll+cedo+al.2002.asynchronous}, propose a
82 version working with integer load. This work was later generalized by
83 the same authors in \cite{cedo+cortes+ripoll+al.2007.convergence}.
84 {\bf Rajouter des choses ici}.
86 Although the Bertsekas and Tsitsiklis' algorithm describes the condition to
87 ensure the convergence, there is no indication or strategy to really implement
88 the load distribution. In other word, a node can send a part of its load to one
89 or many of its neighbors while all the convergence conditions are
90 followed. Consequently, we propose a new strategy called \emph{best effort}
91 that tries to balance the load of a node to all its less loaded neighbors while
92 ensuring that all the nodes concerned by the load balancing phase have the same
93 amount of load. Moreover, when real asynchronous applications are considered,
94 using asynchronous load balancing algorithms can reduce the execution
95 times. Most of the times, it is simpler to distinguish load information messages
96 from data migration messages. Formers ones allows a node to inform its
97 neighbors of its current load. These messages are very small, they can be sent
98 quite often. For example, if an computing iteration takes a significant times
99 (ranging from seconds to minutes), it is possible to send a new load information
100 message at each neighbor at each iteration. Latter messages contains data that
101 migrates from one node to another one. Depending on the application, it may have
102 sense or not that nodes try to balance a part of their load at each computing
103 iteration. But the time to transfer a load message from a node to another one is
104 often much more longer that to time to transfer a load information message. So,
105 when a node receives the information that later it will receive a data message,
106 it can take this information into account and it can consider that its new load
107 is larger. Consequently, it can send a part of it real load to some of its
108 neighbors if required. We call this trick the \emph{virtual load} mechanism.
112 So, in this work, we propose a new strategy for improving the distribution of
113 the load and a simple but efficient trick that also improves the load
114 balancing. Moreover, we have conducted many simulations with SimGrid in order to
115 validate our improvements are really efficient. Our simulations consider that in
116 order to send a message, a latency delays the sending and according to the
117 network performance and the message size, the time of the reception of the
120 In the following of this paper, Section~\ref{BT algo} describes the Bertsekas
121 and Tsitsiklis' asynchronous load balancing algorithm. Moreover, we present a
122 possible problem in the convergence conditions. Section~\ref{Best-effort}
123 presents the best effort strategy which provides an efficient way to reduce the
124 execution times. In Section~\ref{Virtual load}, the virtual load mechanism is
125 proposed. Simulations allowed to show that both our approaches are valid using a
126 quite realistic model detailed in Section~\ref{Simulations}. Finally we give a
127 conclusion and some perspectives to this work.
132 \section{Bertsekas and Tsitsiklis' asynchronous load balancing algorithm}
135 In order prove the convergence of asynchronous iterative load balancing
136 Bertsekas and Tsitsiklis proposed a model
137 in~\cite{bertsekas+tsitsiklis.1997.parallel}. Here we recall some notations.
138 Consider that $N={1,...,n}$ processors are connected through a network.
139 Communication links are represented by a connected undirected graph $G=(N,V)$
140 where $V$ is the set of links connecting different processors. In this work, we
141 consider that processors are homogeneous for sake of simplicity. It is quite
142 easy to tackle the heterogeneous case~\cite{ElsMonPre02}. Load of processor $i$
143 at time $t$ is represented by $x_i(t)\geq 0$. Let $V(i)$ be the set of
144 neighbors of processor $i$. Each processor $i$ has an estimate of the load of
145 each of its neighbors $j \in V(i)$ represented by $x_j^i(t)$. According to
146 asynchronism and communication delays, this estimate may be outdated. We also
147 consider that the load is described by a continuous variable.
149 When a processor send a part of its load to one or some of its neighbors, the
150 transfer takes time to be completed. Let $s_{ij}(t)$ be the amount of load that
151 processor $i$ has transferred to processor $j$ at time $t$ and let $r_{ij}(t)$ be the
152 amount of load received by processor $j$ from processor $i$ at time $t$. Then
153 the amount of load of processor $i$ at time $t+1$ is given by:
155 x_i(t+1)=x_i(t)-\sum_{j\in V(i)} s_{ij}(t) + \sum_{j\in V(i)} r_{ji}(t)
160 Some conditions are required to ensure the convergence. One of them can be
161 called the \emph{ping-pong} condition which specifies that:
163 x_i(t)-\sum _{k\in V(i)} s_{ik}(t) \geq x_j^i(t)+s_{ij}(t)
165 for any processor $i$ and any $j \in V(i)$ such that $x_i(t)>x_j^i(t)$. This
166 condition aims at avoiding a processor to send a part of its load and being
167 less loaded after that.
169 Nevertheless, we think that this condition may lead to deadlocks in some
170 cases. For example, if we consider only three processors and that processor $1$
171 is linked to processor $2$ which is also linked to processor $3$ (i.e. a simple
172 chain which 3 processors). Now consider we have the following values at time $t$:
179 In this case, processor $2$ can either sends load to processor $1$ or processor
180 $3$. If it sends load to processor $1$ it will not satisfy condition
181 (\ref{eq:ping-pong}) because after the sending it will be less loaded that
182 $x_3^2(t)$. So we consider that the \emph{ping-pong} condition is probably to
183 strong. Currently, we did not try to make another convergence proof without this
184 condition or with a weaker condition.
187 \section{Best effort strategy}
190 In this section we describe a new load-balancing strategy that we call
191 \emph{best effort}. The general idea behind this strategy is that each
192 processor, that detects it has more load than some of its neighbors,
193 sends some load to the most of its less loaded neighbors, doing its
194 best to reach the equilibrium between those neighbors and himself.
196 More precisely, when a processor $i$ is in its load-balancing phase,
197 he proceeds as following.
199 \item First, the neighbors are sorted in non-decreasing order of their
200 known loads $x^i_j(t)$.
202 \item Then, this sorted list is traversed in order to find its largest
203 prefix such as the load of each selected neighbor is lesser than:
205 \item the processor's own load, and
206 \item the mean of the loads of the selected neighbors and of the
209 Let's call $S_i(t)$ the set of the selected neighbors, and
210 $\bar{x}(t)$ the mean of the loads of the selected neighbors and of
213 \bar{x}(t) = \frac{1}{\abs{S_i(t)} + 1}
214 \left( x_i(t) + \sum_{j\in S_i(t)} x^i_j(t) \right)
216 The following properties hold:
219 S_i(t) \subset V(i) \\
220 x^i_j(t) < x_i(t) & \forall j \in S_i(t) \\
221 x^i_j(t) < \bar{x} & \forall j \in S_i(t) \\
222 x^i_j(t) \leq x^i_k(t) & \forall j \in S_i(t), \forall k \in V(i) \setminus S_i(t) \\
227 \item Once this selection is completed, processor $i$ sends to each of
228 the selected neighbor $j\in S_i(t)$ an amount of load $s_{ij}(t) =
231 From the above equations, and notably from the definition of
232 $\bar{x}$, it can easily be verified that:
235 x_i(t) - \sum_{j\in S_i(t)} s_{ij}(t) = \bar{x} \\
236 x^i_j(t) + s_{ij}(t) = \bar{x} & \forall j \in S_i(t)
241 \section{Other strategies}
244 \textbf{Question} faut-il décrire les stratégies makhoul et simple ?
246 \paragraph{simple} Tentative de respecter simplement les conditions de Bertsekas.
247 Parmi les voisins moins chargés que soi, on sélectionne :
249 \item un des moins chargés (vmin) ;
250 \item un des plus chargés (vmax),
252 puis on équilibre avec vmin en s'assurant que notre charge reste
253 toujours supérieure à celle de vmin et à celle de vmax.
255 On envoie donc (avec "self" pour soi-même) :
257 \min\left(\frac{load(self) - load(vmin)}{2}, load(self) - load(vmax)\right)
260 \paragraph{makhoul} Ordonne les voisins du moins chargé au plus chargé
261 puis calcule les différences de charge entre soi-même et chacun des
264 Ensuite, pour chaque voisin, dans l'ordre, et tant qu'on reste plus
265 chargé que le voisin en question, on lui envoie 1/(N+1) de la
266 différence calculée au départ, avec N le nombre de voisins.
268 C'est l'algorithme~2 dans~\cite{bahi+giersch+makhoul.2008.scalable}.
270 \section{Virtual load}
273 In this section, we present the concept of \texttt{virtual load}. In order to
274 use this concept, load balancing messages must be sent using two different kinds
275 of messages: load information messages and load balancing messages. More
276 precisely, a node wanting to send a part of its load to one of its neighbors,
277 can first send a load information message containing the load it will send and
278 then it can send the load balancing message containing data to be transferred.
279 Load information message are really short, consequently they will be received
280 very quickly. In opposition, load balancing messages are often bigger and thus
281 require more time to be transferred.
283 The concept of \texttt{virtual load} allows a node that received a load
284 information message to integrate the load that it will receive later in its load
285 (virtually) and consequently send a (real) part of its load to some of its
286 neighbors. In fact, a node that receives a load information message knows that
287 later it will receive the corresponding load balancing message containing the
288 corresponding data. So if this node detects it is too loaded compared to some
289 of its neighbors and if it has enough load (real load), then it can send more
290 load to some of its neighbors without waiting the reception of the load
293 Doing this, we can expect a faster convergence since nodes have a faster
294 information of the load they will receive, so they can take in into account.
296 \textbf{Question} Est ce qu'on donne l'algo avec virtual load?
298 \section{Simulations}
301 In order to test and validate our approaches, we wrote a simulator
303 framework~\cite{casanova+legrand+quinson.2008.simgrid}. This
304 simulator, which consists of about 2,700 lines of C++, allows to run
305 the different load-balancing strategies under various parameters, such
306 as the initial distribution of load, the interconnection topology, the
307 characteristics of the running platform, etc. Then several metrics
308 are issued that permit to compare the strategies.
310 The simulation model is detailed in the next section (\ref{Sim
311 model}), and the experimental contexts are described in
312 section~\ref{Contexts}. Then the results of the simulations are
313 presented in section~\ref{Results}.
315 \subsection{Simulation model}
318 In the simulation model the processors exchange messages which are of
319 two kinds. First, there are \emph{control messages} which only carry
320 information that is exchanged between the processors, such as the
321 current load, or the virtual load transfers if this option is
322 selected. These messages are rather small, and their size is
323 constant. Then, there are \emph{data messages} that carry the real
324 load transferred between the processors. The size of a data message
325 is a function of the amount of load that it carries, and it can be
326 pretty large. In order to receive the messages, each processor has
327 two receiving channels, one for each kind of messages. Finally, when
328 a message is sent or received, this is done by using the non-blocking
329 primitives of SimGrid\footnote{That are \texttt{MSG\_task\_isend()},
330 and \texttt{MSG\_task\_irecv()}.}.
332 During the simulation, each processor concurrently runs three threads:
333 a \emph{receiving thread}, a \emph{computing thread}, and a
334 \emph{load-balancing thread}, which we will briefly describe now.
336 \paragraph{Receiving thread} The receiving thread is in charge of
337 waiting for messages to come, either on the control channel, or on the
338 data channel. Its behavior is sketched by Algorithm~\ref{algo.recv}.
339 When a message is received, it is pushed in a buffer of
340 received message, to be later consumed by one of the other threads.
341 There are two such buffers, one for the control messages, and one for
342 the data messages. The buffers are implemented with a lock-free FIFO
343 \cite{sutter.2008.writing} to avoid contention between the threads.
346 \caption{Receiving thread}
350 \VAR{ctrl\_chan}, \VAR{data\_chan}
351 & communication channels (control and data) \\
352 \VAR{ctrl\_fifo}, \VAR{data\_fifo}
353 & buffers of received messages (control and data) \\
356 wait for a message to be available on either \VAR{ctrl\_chan},
357 or \VAR{data\_chan}\;
358 \If{a message is available on \VAR{ctrl\_chan}}{%
359 get the message from \VAR{ctrl\_chan}, and push it into \VAR{ctrl\_fifo}\;
361 \If{a message is available on \VAR{data\_chan}}{%
362 get the message from \VAR{data\_chan}, and push it into \VAR{data\_fifo}\;
367 \paragraph{Computing thread} The computing thread is in charge of the
368 real load management. As exposed in Algorithm~\ref{algo.comp}, it
369 iteratively runs the following operations:
371 \item if some load was received from the neighbors, get it;
372 \item if there is some load to send to the neighbors, send it;
373 \item run some computation, whose duration is function of the current
374 load of the processor.
376 Practically, after the computation, the computing thread waits for a
377 small amount of time if the iterations are looping too fast (for
378 example, when the current load is near zero).
381 \caption{Computing thread}
385 \VAR{data\_fifo} & buffer of received data messages \\
386 \VAR{real\_load} & current load \\
389 \If{\VAR{data\_fifo} is empty and $\VAR{real\_load} = 0$}{%
390 wait until a message is pushed into \VAR{data\_fifo}\;
392 \While{\VAR{data\_fifo} is not empty}{%
393 pop a message from \VAR{data\_fifo}\;
394 get the load embedded in the message, and add it to \VAR{real\_load}\;
396 \ForEach{neighbor $n$}{%
397 \If{there is some amount of load $a$ to send to $n$}{%
398 send $a$ units of load to $n$, and subtract it from \VAR{real\_load}\;
401 \If{$\VAR{real\_load} > 0.0$}{
402 simulate some computation, whose duration is function of \VAR{real\_load}\;
403 ensure that the main loop does not iterate too fast\;
408 \paragraph{Load-balancing thread} The load-balancing thread is in
409 charge of running the load-balancing algorithm, and exchange the
410 control messages. It iteratively runs the following operations:
412 \item get the control messages that were received from the neighbors;
413 \item run the load-balancing algorithm;
414 \item send control messages to the neighbors, to inform them of the
415 processor's current load, and possibly of virtual load transfers;
416 \item wait a minimum (configurable) amount of time, to avoid to
421 \caption{Load-balancing}
424 \While{\VAR{ctrl\_fifo} is not empty}{%
425 pop a message from \VAR{ctrl\_fifo}\;
426 identify the sender of the message,
427 and update the current knowledge of its load\;
429 run the load-balancing algorithm to make the decision about load transfers\;
430 \ForEach{neighbor $n$}{%
431 send a control messages to $n$\;
433 ensure that the main loop does not iterate too fast\;
438 For the sake of simplicity, a few details were voluntary omitted from
439 these descriptions. For an exhaustive presentation, we refer to the
440 actual code that was used for the experiments, and which is
441 available at \textbf{FIXME URL}.
443 \textbf{FIXME: ajouter des détails sur la gestion de la charge virtuelle ?}
445 \subsection{Experimental contexts}
448 \paragraph{Configurations}
450 \item[\textbf{platforms}] homogeneous (cluster); heterogeneous (subset
452 \item[\textbf{platform size}] platforms with 16, 64, 256, and 1024 nodes
453 \item[\textbf{topologies}] line; torus; hypercube
454 \item[\textbf{initial load distribution}] initially on a only node;
455 initially on all nodes
456 \item[\textbf{comp/comm ratio}] $10/1$, $1/1$, $1/10$
459 \paragraph{Algorithms}
461 \item[\textbf{strategies}] makhoul; besteffort with $k\in \{1,2,4\}$
462 \item[\textbf{variants}] with, and without virtual load (bookkeeping)
463 \item[\textbf{domain}] real load, and integer load
469 \item[\textbf{average idle time}]
470 \item[\textbf{average convergence date}]
471 \item[\textbf{maximum convergence date}]
472 \item[\textbf{data transfer amount}] relative to the total data amount
475 \subsection{Validation of our approaches}
479 On veut montrer quoi ? :
481 1) best plus rapide que les autres (simple, makhoul)
482 2) avantage virtual load
484 Est ce qu'on peut trouver des contre exemple?
488 Simulation avec temps définies assez long et on mesure la qualité avec : volume de calcul effectué, volume de données échangées
489 Mais aussi simulation avec temps court qui montre que seul best converge
492 Expés avec ratio calcul/comm rapide et lent
494 Quelques expés avec charge initiale aléatoire plutot que sur le premier proc
496 Cadre processeurs homogènes
500 On ne tient pas compte de la vitesse des liens donc on la considère homogène
502 Prendre un réseau hétérogène et rendre processeur homogène
504 Taille : 10 100 très gros
506 \section{Conclusion and perspectives}
509 \bibliographystyle{spmpsci}
510 \bibliography{biblio}
517 %%% ispell-local-dictionary: "american"
520 % LocalWords: Raphaël Couturier Arnaud Giersch Abderrahmane Sider Franche ij
521 % LocalWords: Bertsekas Tsitsiklis SimGrid DASUD Comté Béjaïa asynchronism ji
522 % LocalWords: ik isend irecv Cortés et al chan ctrl fifo