comparison writeup/nipswp_submission.tex @ 598:a0fdc1f134da

minor changes to nips workshop submission
author Yoshua Bengio <bengioy@iro.umontreal.ca>
date Thu, 14 Oct 2010 22:09:55 -0400
parents 5ab605c9a7d9
children
comparison
equal deleted inserted replaced
597:5ab605c9a7d9 598:a0fdc1f134da
172 which is based on training with or without these transformed images and testing on 172 which is based on training with or without these transformed images and testing on
173 clean ones. We measure the relative advantage of out-of-distribution examples 173 clean ones. We measure the relative advantage of out-of-distribution examples
174 (perturbed or out-of-class) 174 (perturbed or out-of-class)
175 for a deep learner vs a supervised shallow one. 175 for a deep learner vs a supervised shallow one.
176 Code for generating these transformations as well as for the deep learning 176 Code for generating these transformations as well as for the deep learning
177 algorithms are made available at {\tt http://hg.assembla.com/ift6266}. 177 algorithms are made available at {\tt http://anonymous.url.net}.%{\tt http://hg.assembla.com/ift6266}.
178 We also estimate the relative advantage for deep learners of training with 178 We also estimate the relative advantage for deep learners of training with
179 other classes than those of interest, by comparing learners trained with 179 other classes than those of interest, by comparing learners trained with
180 62 classes with learners trained with only a subset (on which they 180 62 classes with learners trained with only a subset (on which they
181 are then tested). 181 are then tested).
182 The conclusion discusses 182 The conclusion discusses
225 \subfigure[Gaussian Noise]{\includegraphics[scale=0.6]{images/Distorsiongauss_only.png}} 225 \subfigure[Gaussian Noise]{\includegraphics[scale=0.6]{images/Distorsiongauss_only.png}}
226 \subfigure[Background Image Addition]{\includegraphics[scale=0.6]{images/background_other_only.png}} 226 \subfigure[Background Image Addition]{\includegraphics[scale=0.6]{images/background_other_only.png}}
227 \subfigure[Salt \& Pepper]{\includegraphics[scale=0.6]{images/Poivresel_only.png}} 227 \subfigure[Salt \& Pepper]{\includegraphics[scale=0.6]{images/Poivresel_only.png}}
228 \subfigure[Scratches]{\includegraphics[scale=0.6]{images/Rature_only.png}} 228 \subfigure[Scratches]{\includegraphics[scale=0.6]{images/Rature_only.png}}
229 \subfigure[Grey Level \& Contrast]{\includegraphics[scale=0.6]{images/Contrast_only.png}} 229 \subfigure[Grey Level \& Contrast]{\includegraphics[scale=0.6]{images/Contrast_only.png}}
230 \caption{Transformation modules} 230 \caption{Top left (a): example original image. Others (b-o): examples of the effect
231 of each transformation module taken separately. Actual perturbed examples are obtained by
232 a pipeline of these, with random choices about which module to apply and how much perturbation
233 to apply.}
231 \label{fig:transform} 234 \label{fig:transform}
232 \vspace*{-2mm} 235 \vspace*{-2mm}
233 \end{figure} 236 \end{figure}
234 237
235 \vspace*{-3mm} 238 \vspace*{-3mm}
245 248
246 The first step in constructing the larger datasets (called NISTP and P07) is to sample from 249 The first step in constructing the larger datasets (called NISTP and P07) is to sample from
247 a {\em data source}: {\bf NIST} (NIST database 19), {\bf Fonts}, {\bf Captchas}, 250 a {\em data source}: {\bf NIST} (NIST database 19), {\bf Fonts}, {\bf Captchas},
248 and {\bf OCR data} (scanned machine printed characters). Once a character 251 and {\bf OCR data} (scanned machine printed characters). Once a character
249 is sampled from one of these sources (chosen randomly), the second step is to 252 is sampled from one of these sources (chosen randomly), the second step is to
250 apply a pipeline of transformations and/or noise processes described in section \ref{s:perturbations}. 253 apply a pipeline of transformations and/or noise processes outlined in section \ref{s:perturbations}.
251 254
252 To provide a baseline of error rate comparison we also estimate human performance 255 To provide a baseline of error rate comparison we also estimate human performance
253 on both the 62-class task and the 10-class digits task. 256 on both the 62-class task and the 10-class digits task.
254 We compare the best Multi-Layer Perceptrons (MLP) against 257 We compare the best Multi-Layer Perceptrons (MLP) against
255 the best Stacked Denoising Auto-encoders (SDA), when 258 the best Stacked Denoising Auto-encoders (SDA), when
256 both models' hyper-parameters are selected to minimize the validation set error. 259 both models' hyper-parameters are selected to minimize the validation set error.
257 We also provide a comparison against a precise estimate 260 We also provide a comparison against a precise estimate
258 of human performance obtained via Amazon's Mechanical Turk (AMT) 261 of human performance obtained via Amazon's Mechanical Turk (AMT)
259 service (http://mturk.com). 262 service ({\tt http://mturk.com}).
260 AMT users are paid small amounts 263 AMT users are paid small amounts
261 of money to perform tasks for which human intelligence is required. 264 of money to perform tasks for which human intelligence is required.
262 Mechanical Turk has been used extensively in natural language processing and vision. 265 Mechanical Turk has been used extensively in natural language processing and vision.
263 %processing \citep{SnowEtAl2008} and vision 266 %processing \citep{SnowEtAl2008} and vision
264 %\citep{SorokinAndForsyth2008,whitehill09}. 267 %\citep{SorokinAndForsyth2008,whitehill09}.