Mercurial > ift6266
diff writeup/nips2010_submission.tex @ 466:6205481bf33f
asking the questions
author | Yoshua Bengio <bengioy@iro.umontreal.ca> |
---|---|
date | Fri, 28 May 2010 17:39:22 -0600 |
parents | 24f4a8b53fcc |
children | e0e57270b2af |
line wrap: on
line diff
--- a/writeup/nips2010_submission.tex Fri May 28 17:33:15 2010 -0600 +++ b/writeup/nips2010_submission.tex Fri May 28 17:39:22 2010 -0600 @@ -89,6 +89,22 @@ converted into a deep supervised feedforward neural network and trained by stochastic gradient descent. +In this paper we ask the following questions: +\begin{enumerate} +\item Do the good results previously obtained with deep architectures on the +MNIST digits generalize to the setting of a much larger and richer (but similar) +dataset, the NIST special database 19, with 62 classes and around 800k examples? +\item To what extent does the perturbation of input images (e.g. adding +noise, affine transformations, background images) make the resulting +classifier better not only on similarly perturbed images but also on +the {\em original clean examples}? +\item Do deep architectures benefit more from such {\em out-of-distribution} +examples, i.e. do they benefit more from the self-taught learning~\cite{RainaR2007} framework? +\item Similarly, does the feature learning step in deep learning algorithms benefit more +training with similar but different classes (i.e. a multi-task learning scenario) than +a corresponding shallow and purely supervised architecture? +\end{enumerate} +The experimental results presented here provide positive evidence towards all of these questions. \section{Perturbation and Transformation of Character Images}