# HG changeset patch # User Yoshua Bengio # Date 1275089962 21600 # Node ID 6205481bf33f2a98e24ff286432e4e6b0771d38a # Parent a48601e8d431ca441ab3de4363f38c32b9b9ca38 asking the questions diff -r a48601e8d431 -r 6205481bf33f writeup/nips2010_submission.tex --- a/writeup/nips2010_submission.tex Fri May 28 17:33:15 2010 -0600 +++ b/writeup/nips2010_submission.tex Fri May 28 17:39:22 2010 -0600 @@ -89,6 +89,22 @@ converted into a deep supervised feedforward neural network and trained by stochastic gradient descent. +In this paper we ask the following questions: +\begin{enumerate} +\item Do the good results previously obtained with deep architectures on the +MNIST digits generalize to the setting of a much larger and richer (but similar) +dataset, the NIST special database 19, with 62 classes and around 800k examples? +\item To what extent does the perturbation of input images (e.g. adding +noise, affine transformations, background images) make the resulting +classifier better not only on similarly perturbed images but also on +the {\em original clean examples}? +\item Do deep architectures benefit more from such {\em out-of-distribution} +examples, i.e. do they benefit more from the self-taught learning~\cite{RainaR2007} framework? +\item Similarly, does the feature learning step in deep learning algorithms benefit more +training with similar but different classes (i.e. a multi-task learning scenario) than +a corresponding shallow and purely supervised architecture? +\end{enumerate} +The experimental results presented here provide positive evidence towards all of these questions. \section{Perturbation and Transformation of Character Images}