-\begin{figure}
+\begin{figure*}
\begin{center}
-\includegraphics[scale=0.3]{reseau.png}
+\includegraphics[scale=0.2]{SensorNetwork.png}
\begin{scriptsize}
-An example of a sensor network ofsize 10. All nodes are video sensor
-except the 5 and the 9 one which is the sink.
-\JFC{reprendre la figure, trouver un autre titre}
-\end{scriptsize}
-\caption{SN with 10 sensor}\label{fig:sn}.
+An example of a sensor network of size 10.
+All nodes are video sensors (depicted as small discs)
+except the 9 one which is the sink (depicted as a rectangle).
+Large lircles represent the maximum
+transmission range which is set to 20 in a square region which is
+$50 m \times 50 m$.
+\end{scriptsize}
+\caption{Illustration of a SN of size 10}\label{fig:sn}.
\end{center}
-\end{figure}
+\end{figure*}
Let us first recall the basics of the~\cite{HLG09} article.
The video sensor network is memorized as a connected non oriented
-oriented labelled graph.
+graph.
In this one,
the nodes, in a set $N$, are sensors, links, or the sink.
Furthermore, there is an edge from $i$ to $j$ if $i$ can
-send a message to $j$. The set of all edges is further denoted as
+send a message to $j$, \textit{i. e.}, the distance betwween
+$i$ and $j$ is less than a given maximum
+transmission range.
+All the possible edges are stored into a sequence
$L$.
Figure~\ref{fig:sn} gives an example of such a network.
where
$a_{il}$ is 1 if $l$ starts with $i$, is -1 if $l$ ends width $i$
and 0 otherwise.
-
+Moreover, the outgoing links(resp. the incoming links) are represented
+with the $A^+$ matrix (res. with the $A^-$ matrix), whose elements are defined:
+$a_{il}^+$ (resp. $a_{il}^-$) is 1 if the link $l$ is an outgoing link from $i$
+(resp an incoming link into $i$) and 0 otherwise.
Let $V \subset N $ be the set of the video sensors of $N$.
Let thus $R_h$, $R_h \geq 0$,
be the encoding rate of video sensor $h$, $h \in V$.
-Let $\eta_{hi}$ be the production rate of the node $i$,
-for the session initiated by $h$. More precisely, we have
+Let $\eta_{hi}$ be the rate inside the node $i$
+of the production that has beeninitiated by $h$. More precisely, we have
$ \eta_{hi}$ is equal to $ R_h$ if $i$ is $h$,
is equal to $-R_h$ if $i$ is the sink, and $0$ otherwise.
-We are then left to focus on the flows in this network.
+Let us focus on the flows in this network.
Let $x_{hl}$, $x_{hl}\geq 0$, be the flow inside the edge $l$ that
-issued from the $h$ session and
+issued from the node $h$ and
let $y_l = \sum_{h \in V}x_{hl} $ the sum of all the flows inside $l$.
Thus, what is produced inside the $i^{th}$ sensor for session $h$
is $ \eta_{hi} = \sum_{l \in L }a_{il}x_{hl} $.
The encoding power of the $i$ node is $P_{si}$, $P_{si} > 0$.
-
-The distortion is bounded $\sigma^2 e^{-\gamma . R_h.P_{sh}^{}2/3} \leq D_h$.
-
+The emmission distortion of the $i$ node is
+$\sigma^2 e^{-\gamma . R_i.P_{si}^{}2/3}$
+where $\sigma^2$ is the average input variance and
+$\gamma$ is the encoding efficiency coefficient.
+This distortion
+is bounded by a constant value $D_h$.
The initial energy of the $i$ node is $B_i$.
-
+The transmission consumed power of node $i$ is
+$P_{ti} = c_l^s.y_l$ where $c_l^s$ is the transmission energy
+consumption cost of link $l$, $l\in L$. This cost is defined
+as foolows: $c_l^s = \alpha +\beta.d_l^{n_p} $ wehre
+$d_l$ represents the distance of the link $l$,
+$\alpha$, $\beta$, and $n_p$ are constant.
+The reception consumed power of node $i$ is
+$P_{ri} = c^r \sum_{l \in L } a_{il}^-.y_l$
+where $c^r$ is a reception energy consumption cost.
The overall consumed power of the $i$ node is
$P_{si}+ P_{ti} + P_{ri}=
P_{si}+ \sum_{l \in L}a_{il}^{+}.c^s_l.y_l +
\item $P_{sh} > 0,\forall h \in V$
\end{enumerate}
-
To achieve this optimizing goal
a local optimisation, the problem is translated into an
equivalent one: find $R$, $x$, $P_s$ which minimize
\end{array}
$$
-The authors then apply a dual based approach with Lagrange multiplier
-to solve such a problem.
-They first introduce dual variables
-$u_{hi}$, $v_{h}$, $\lambda_{i}$, and $w_l$ for any
-$h \in V$,$ i \in N$, and $l \in L$.
-They next replace the objective of reducing $\sum_{i \in N }q_i^2$
+
+They thus replace the objective of reducing
+$\sum_{i \in N }q_i^2$
by the objective of reducing
-$$
+\begin{equation}
\sum_{i \in N }q_i^2 +
\sum_{h \in V, l \in L } \delta.x_{hl}^2
+ \sum_{h \in V }\delta.R_{h}^2
-$$
-where $\delta$ is a \JFC{ formalisation} factor.
-This allows indeed to get convex functions whose minimum value is unique.
+\label{eq:obj2}
+\end{equation}
+where $\delta$ is a regularisation factor.
+This indeed introduces quadratic functions on variables $x_{hl}$ and
+$R_{h}$ and makes some of the functions strictly convex.
+
+The authors then apply a classical dual based approach with Lagrange multiplier
+to solve such a problem~\cite{}.
+They first introduce dual variables
+$u_{hi}$, $v_{h}$, $\lambda_{i}$, and $w_l$ for any
+$h \in V$,$ i \in N$, and $l \in L$.
-The proposed algorithm iteratively computes the following variables
+\begin{equation}
+\begin{array}{l}
+L(R,x,P_{s},q,u,v,\lambda,w)=\\
+\sum_{i \in N} q_i^2 + q_i. \left(
+\sum_{l \in L } a_{il}w_l^{(k)}-
+\lambda_iB_i
+\right)\\
++ \sum_{h \in V}
+v_h.\dfrac{\ln(\sigma^2/D_h)}{\gamma P_{sh} ^{2/3}} + \lambda_h P_{sh}\\
++ \sum_{h \in V} \sum_{l\in L}
+\left(
+\delta.x_{hl}^2 \right.\\
+\qquad \qquad + x_{hl}.
+\sum_{i \in N} \left(
+\lambda_{i}.(c^s_l.a_{il}^{+} +
+c^r. a_{il}^{-} ) \right.\\
+\qquad \qquad\qquad \qquad +
+\left.\left. u_{hi} a_{il}
+\right)
+\right)\\
+ + \sum_{h \in V}
+\delta R_{h}^2
+-v_h.R_{h} - \sum_{i \in N} u_{hi}\eta_{hi}
+\end{array}
+\end{equation}
+The proposed algorithm iteratively computes the following variables
+until the variation of the dual function is less than a given threshold.
\begin{enumerate}
\item $ u_{hi}^{(k+1)} = u_{hi}^{(k)} - \theta^{(k)}. \left(
\eta_{hi}^{(k)} - \sum_{l \in L }a_{il}x_{hl}^{(k)} \right) $
$v_{h}^{(k+1)}= \max\left\{0,v_{h}^{(k)} - \theta^{(k)}.\left( R_h^{(k)} - \dfrac{\ln(\sigma^2/D_h)}{\gamma.(P_{sh}^{(k)})^{2/3}} \right)\right\}$
\item
$\begin{array}{l}
- \lambda_{i}^{(k+1)} = \lambda_{i}^{(k)} - \theta^{(k)}.\left(
- q^{(k)}.B_i \right.\\
+ \lambda_{i}^{(k+1)} = \max\left\{0, \lambda_{i}^{(k)} - \theta^{(k)}.\left(
+ q^{(k)}.B_i \right. \right.\\
\qquad\qquad\qquad -\sum_{l \in L}a_{il}^{+}.c^s_l.\left( \sum_{h \in V}x_{hl}^{(k)} \right) \\
- \qquad\qquad\qquad - \left. \sum_{l \in L} a_{il}^{-}.c^r.\left( \sum_{h \in V}x_{hl}^{(k)} \right) - P_{si}^{(k)} \right)
+ \qquad\qquad\qquad - \left.\left. \sum_{l \in L} a_{il}^{-}.c^r.\left( \sum_{h \in V}x_{hl}^{(k)} \right) - P_{si}^{(k)} \right) \right\}
\end{array}
$
\item
$q_i^{(k)} = \arg\min_{q_i>0}
\left(
-q^2 + q.
+q_i^2 + q_i.
\left(
\sum_{l \in L } a_{il}w_l^{(k)}-
\lambda_i^{(k)}B_i
\right)
\right)$
-\item
+\item \label{item:psh}
$
P_{sh}^{(k)}
=
\arg \min_{p > 0}
\left(
-v_h^{(k)}.\dfrac{\ln(\sigma^2/D_h)}{\gamma p ^{2/3}} + \lambda_h^{(k)}p
+v_h^{(k)}.\dfrac{\ln(\sigma^2/D_h)}{\gamma p^{2/3}} + \lambda_h^{(k)}p
\right)
$
$
\end{enumerate}
where the first four elements are dual variable and the last four ones are
-primal ones
\ No newline at end of file
+primal ones.
\ No newline at end of file