From: couchot Date: Tue, 15 Jan 2013 13:18:41 +0000 (+0100) Subject: comparaison avec HUGO et autres X-Git-Url: https://bilbo.iut-bm.univ-fcomte.fr/and/gitweb/canny.git/commitdiff_plain/51c827b374614757c68f7d7cd8d799eb46f477e0?ds=sidebyside comparaison avec HUGO et autres --- diff --git a/biblio.bib b/biblio.bib index 3c06dc7..89cddc2 100644 --- a/biblio.bib +++ b/biblio.bib @@ -178,7 +178,7 @@ author = {Jessica J. Fridrich and @inproceedings{DBLP:conf/ih/PevnyFB10, added-at = {2010-10-08T00:00:00.000+0200}, - author = {Pevný, Tomás and Filler, Tomás and Bas, Patrick}, + author = {Pevn{\'y}, Tom{\'a}s and Filler, Tom{\'a}s and Bas, Patrick}, biburl = {http://www.bibsonomy.org/bibtex/28d83b7eac2c22ed5e7e072fd43a34248/dblp}, booktitle = {Information Hiding}, crossref = {DBLP:conf/ih/2010}, @@ -198,6 +198,18 @@ author = {Jessica J. Fridrich and year = 2010 } + +@Misc{Boss10, + OPTkey = {}, + author = {Pevný, Tomáš and Filler, Tomáš and Bas, Patrick}, + title = {Break Our Steganographic System}, + OPThowpublished = {}, + OPTmonth = {}, + year = {2010}, + note = { available at \url{http://www.agents.cz/boss/}}, + OPTannote = {} +} + @proceedings{DBLP:conf/ih/2010, editor = {Rainer B{\"o}hme and Philip W. L. Fong and @@ -321,7 +333,7 @@ author = {Jessica J. Fridrich and @inproceedings{DBLP:conf/mmsec/FridrichPK07, added-at = {2007-10-26T00:00:00.000+0200}, - author = {Fridrich, Jessica J. and Pevný, Tomás and Kodovský, Jan}, + author = {Fridrich, Jessica J. and Pevn{\'y}, Tom{\'a}s and Kodovsk{\'y}, Jan}, biburl = {http://www.bibsonomy.org/bibtex/26123ac512b5e1fe72a44e73d101d8b95/dblp}, booktitle = {MMSec}, crossref = {DBLP:conf/mmsec/2007}, @@ -597,7 +609,7 @@ author = {Jessica J. Fridrich and @InProceedings{KF11, - author = {Jan Kodovský and Jessica Fridrich}, + author = {Jan Kodovsk{\'y} and Jessica Fridrich}, title = {Steganalysis in high dimensions: Fusing classifiers built on random subspaces}, OPTcrossref = {}, OPTkey = {}, diff --git a/experiments.tex b/experiments.tex index e4db410..60cce0b 100644 --- a/experiments.tex +++ b/experiments.tex @@ -1,16 +1,35 @@ +For the whole experiment, a set of 500 images is randomly extracted +from the database taken from the BOSS contest~\cite{Boss10}. +In this set, each cover is a $512\times 512$ +grayscale digital image. + + \subsection{Adaptive Embedding Rate} Two strategies have been developed in our scheme with respect to the rate of -embedding which is either \emph{ adaptive} or \emph{fixed}. +embedding which is either \emph{adaptive} or \emph{fixed}. In the former the embedding rate depends on the number of edge pixels. The higher it is, the larger is the message length that can be considered. Practically, a set of edge pixels is computed according to the Canny algorithm with high threshold. The message length is thus defined to be the half of this set cardinality. -The rate between available bits and bit message length is then more than two.This constraint is indeed induced by the fact that the efficiency +In this strategy, two methods are thus applied to extract bits that +are modified. The first one is a direct application of the STC algorithm. +This method is further refered as \emph{adaptive+STC}. +The second one randomly choose the subset of pixels to modify by +applying the BBS PRNG again. This method is denoted \emph{adaptive+sample}. +Notice that the rate between +available bits and bit message length is always equal to two. +This constraint is indeed induced by the fact that the efficiency of the STC algorithm is unsatisfactory under that threshold. +On our experiments and with the adaptive scheme, +the average size of the message that can be embedded is 16445. +Its corresponds to an average payload of 6.35\%. + + + In the latter, the embedding rate is defined as a percentage between the number of the modified pixels and the length of the bit message. @@ -19,9 +38,7 @@ Practically, the Canny algorithm generates a a set of edge pixels with threshold that is decreasing until its cardinality is sufficient. If the set cardinality is more than twice larger than the bit message length an STC step is again applied. -Otherwise, pixels are randomly chosen from the set of pixels to build the -subset with a given size. The BBS PRNG is again applied there. - +Otherwise, pixels are again randomly chosen with BBS. @@ -38,28 +55,46 @@ The other last ones have been designed to tackle this problem. \begin{table} \begin{center} -\begin{tabular}{|c|c|c|} +\begin{tabular}{|c|c|c||c|c|} \hline -Embedding rate & Adaptive & 10 \% \\ + & \multicolumn{2}{|c||}{Adaptive} & fixed & HUGO \\ +Embedding rate & + STC & + sample & 10\% & 10\%\\ \hline -PSNR & 66.55 & 61.86 \\ +PSNR & 66.55 & 63.48 & 61.86 & 64.65 \\ \hline -PSNR-HVS-M & 78.6 & 72.9 \\ +PSNR-HVS-M & 78.6 & 75.39 & 72.9 & 76.67\\ \hline -BIQI & 28.3 & 28.4 \\ +BIQI & 28.3 & 28.28 & 28.4 & 28.28\\ \hline -wPSNR & 86.43& 77.47 \\ +wPSNR & 86.43& 80.59 & 77.47& 83.03\\ \hline \end{tabular} \end{center} \caption{Quality measures of our steganography approach\label{table:quality}} \end{table} - -Let us compare the STABYLO approach with other edge based steganography +Let us give an interpretation of these experiments. +First of all, the adaptive strategy produces images with lower distortion +than the one of images resulting from the 10\% fixed strategy. +Numerical results are indeed always greater for the former strategy than +for the latter, except for the BIQI metrics where differences are not relevent. +These results are not surprising since the adaptive strategy aims at +embedding messages whose length is decided according to a higher threshold +into the edge detection. +Let us focus on the quality of HUGO images: with a given fixed +embedding rate (10\%) +HUGO always produces images whose quality is higher than the STABYLO's one. +However, our approach nevertheless provides beter results with the strategy +adaptive+STC in a lightweight manner, as motivated in the introduction. + + +Let us now compare the STABYLO approach with other edge based steganography schemes with respect to the image quality. -First of all, wPSNR and PSNR of the Edge Adaptive -scheme detailed in~\cite{Luo:2010:EAI:1824719.1824720} are lower than ours. +First of all, the Edge Adaptive +scheme detailed in~\cite{Luo:2010:EAI:1824719.1824720} +executed with a 10\% embedding rate +has the same PSNR but a lower wPSNR than our: +these two metrics are respectively equal to 61.9 and 68.9. Next both the approaches~\cite{DBLP:journals/eswa/ChenCL10,Chang20101286} focus on increasing the payload while the PSNR is acceptable, but do not give quality metrics for fixed embedding rate from a large base of images. @@ -67,6 +102,8 @@ Our approach outperforms the former thanks to the introduction of the STC algorithm. + + \subsection{Steganalysis} @@ -89,15 +126,17 @@ can be a favourably executed thanks to an Ensemble Classifiers. \begin{table} \begin{center} -\begin{tabular}{|c|c|c|c|} +\begin{tabular}{|c|c|c|c|c|} \hline -Schemes & \multicolumn{2}{|c|}{STABYLO} & HUGO\\ +Schemes & \multicolumn{3}{|c|}{STABYLO} & HUGO\\ \hline -Embedding rate & Adaptive & 10 \% & 10 \%\\ +Embedding rate & \multicolumn{2}{|c|}{Adaptive} & 10 \% & 10 \%\\ + & + STC & + sample & & \\ + \hline -AUMP & 0.39 & 0.22 & 0.50 \\ +AUMP & 0.39 & & 0.22 & 0.50 \\ \hline -Ensemble Classifier & 0.47 & 0.35 & 0.48 \\ +Ensemble Classifier & 0.47 & & 0.35 & 0.48 \\ \hline \end{tabular} diff --git a/stc/exp/raphus/test_wpsnr.py b/stc/exp/raphus/test_wpsnr.py index f735467..20db963 100644 --- a/stc/exp/raphus/test_wpsnr.py +++ b/stc/exp/raphus/test_wpsnr.py @@ -1,6 +1,6 @@ import Image as im import numpy as np -from Image import ImageStat as imst +#from Image import ImageStat as imst from numpy import linalg as LA from math import *