%First, performances of the Ehrlich-Aberth method of root finding polynomials
%implemented on CPUs and on GPUs are studied.
%First, performances of the Ehrlich-Aberth method of root finding polynomials
%implemented on CPUs and on GPUs are studied.
-We performed a set of experiments on the sequential and the parallel algorithms, for both sparse and full polynomials and different sizes. We took into account the execution times, the polynomial size and the number of threads per block performed by sum or each experiment on CPUs and on GPUs.
+We performed a set of experiments on the sequential and the parallel algorithms, for both sparse and full polynomials and different sizes. We took into account the execution times, the polynomial size and the number of threads per block performed by sum or each experiment on CPU and on GPU.
All experimental results obtained from the simulations are made in
double precision data, the convergence threshold of the methods is set
All experimental results obtained from the simulations are made in
double precision data, the convergence threshold of the methods is set
\subsection{Comparison of execution times of the Ehrlich-Aberth method
on a CPU with OpenMP (1 core and 4 cores) vs. on a Tesla GPU}
\subsection{Comparison of execution times of the Ehrlich-Aberth method
on a CPU with OpenMP (1 core and 4 cores) vs. on a Tesla GPU}
\subsection{Influence of the number of threads on the execution times of different polynomials (sparse and full)}
To optimize the performances of an algorithm on a GPU, it is necessary to maximize the use of cores GPU (maximize the number of threads executed in parallel) and to optimize the use of the various memoirs GPU. In fact, it is interesting to see the influence of the number of threads per block on the execution time of Ehrlich-Aberth algorithm.
\subsection{Influence of the number of threads on the execution times of different polynomials (sparse and full)}
To optimize the performances of an algorithm on a GPU, it is necessary to maximize the use of cores GPU (maximize the number of threads executed in parallel) and to optimize the use of the various memoirs GPU. In fact, it is interesting to see the influence of the number of threads per block on the execution time of Ehrlich-Aberth algorithm.
-For that, we notice that the maximum number of threads per block for the Nvidia Tesla K40 GPU is 1024, so we varied the number of threads per block from 8 to 1024. We took into account the execution time for both sparse and full of 10 different polynomials of size 50,000 and 10 different polynomials of size 500,000 degrees.
+For that, we notice that the maximum number of threads per block for the Nvidia Tesla K40 GPU is 1024, so we varied the number of threads per block from 8 to 1,024. We took into account the execution time for both sparse and full of 10 different polynomials of size 50,000 and 10 different polynomials of size 500,000 degrees.
execution time of the Ehrlich-Aberth algorithm without this solution,
with full and sparse polynomials degrees. We can see that the
execution times for both algorithms are the same with full polynomials
execution time of the Ehrlich-Aberth algorithm without this solution,
with full and sparse polynomials degrees. We can see that the
execution times for both algorithms are the same with full polynomials
also clearly show that the classical version (without log.exp) of
Ehrlich-Aberth algorithm do not converge after these degree with
sparse and full polynomials. In counterpart, the new version of
also clearly show that the classical version (without log.exp) of
Ehrlich-Aberth algorithm do not converge after these degree with
sparse and full polynomials. In counterpart, the new version of
high degree polynomials.
%in fact, when the modulus of the roots are up than \textit{R} given in ~\ref{R},this exceed the limited number in the mantissa of floating points representations and can not compute the iterative function given in ~\ref{eq:Aberth-H-GS} to obtain the root solution, who justify the divergence of the classical Ehrlich-Aberth algorithm. However, applying log.exp solution given in ~\ref{sec2} took into account the limit of floating using the iterative function in(Eq.~\ref{Log_H1},Eq.~\ref{Log_H2} and allows to solve a very large polynomials degrees .
high degree polynomials.
%in fact, when the modulus of the roots are up than \textit{R} given in ~\ref{R},this exceed the limited number in the mantissa of floating points representations and can not compute the iterative function given in ~\ref{eq:Aberth-H-GS} to obtain the root solution, who justify the divergence of the classical Ehrlich-Aberth algorithm. However, applying log.exp solution given in ~\ref{sec2} took into account the limit of floating using the iterative function in(Eq.~\ref{Log_H1},Eq.~\ref{Log_H2} and allows to solve a very large polynomials degrees .
sparse polynomial degrees ranging from 1,000 to 1,000,000. We can see
that the Ehrlich-Aberth algorithm is faster than Durand-Kerner
algorithm, with an average of 25 times faster. Then, when degrees of
sparse polynomial degrees ranging from 1,000 to 1,000,000. We can see
that the Ehrlich-Aberth algorithm is faster than Durand-Kerner
algorithm, with an average of 25 times faster. Then, when degrees of
-polynomial exceed 500000 the execution time with EA is of the order
-100 whereas DK passes in the order 1000.
+polynomial exceed 500,000 the execution time with EA is of the order
+100 whereas DK passes in the order 1,000.