diff writeup/aistats2011_submission.tex @ 603:eb6244c6d861

aistats submission
author Yoshua Bengio <bengioy@iro.umontreal.ca>
date Sun, 31 Oct 2010 22:40:33 -0400
parents 203c6071e104
children 51213beaed8b
line wrap: on
line diff
--- a/writeup/aistats2011_submission.tex	Sun Oct 31 22:27:30 2010 -0400
+++ b/writeup/aistats2011_submission.tex	Sun Oct 31 22:40:33 2010 -0400
@@ -109,7 +109,7 @@
 stochastic gradient descent.
 One of these layer initialization techniques,
 applied here, is the Denoising
-Auto-encoder~(DAE)~\citep{VincentPLarochelleH2008-very-small} (see
+Auto-encoder~(DA)~\citep{VincentPLarochelleH2008-very-small} (see
 Figure~\ref{fig:da}), which performed similarly or 
 better~\citep{VincentPLarochelleH2008-very-small} than previously
 proposed Restricted Boltzmann Machines (RBM)~\citep{Hinton06} 
@@ -118,10 +118,10 @@
 to denoise its input, creating a layer of features that can be used as
 input for the next layer. Note that training a Denoising Auto-Encoder
 can actually been seen as training a particular RBM by an inductive
-principle different from maximum likelihood~\cite{Vincent-SM-2010}, namely by
-Score Matching~\citep{Hyvarinen-2005,HyvarinenA2008}. 
+principle different from maximum likelihood~\citep{ift6266-tr-anonymous}, % Vincent-SM-2010}, 
+namely by Score Matching~\citep{Hyvarinen-2005,HyvarinenA2008}. 
 
-Previous comparative experimental results with stacking of RBMs and DAEs
+Previous comparative experimental results with stacking of RBMs and DAs
 to build deep supervised predictors had shown that they could outperform
 shallow architectures in a variety of settings (see~\citet{Bengio-2009}
 for a review), especially
@@ -161,9 +161,9 @@
 
 %
 The {\bf main claim} of this paper is that deep learners (with several levels of representation) can
-{\bf benefit more from self-taught learning than shallow learners} (with a single
-level), both in the context of the multi-task setting and from {\em
-  out-of-distribution examples} in general. Because we are able to improve on state-of-the-art
+{\bf benefit more from out-of-distribution examples than shallow learners} (with a single
+level), both in the context of the multi-task setting and from
+ perturbed examples. Because we are able to improve on state-of-the-art
 performance and reach human-level performance
 on a large-scale task, we consider that this paper is also a contribution
 to advance the application of machine learning to handwritten character recognition.
@@ -212,7 +212,7 @@
 are then tested).
 The conclusion discusses
 the more general question of why deep learners may benefit so much from 
-the self-taught learning framework. Since out-of-distribution data
+out-of-distribution examples. Since out-of-distribution data
 (perturbed or from other related classes) is very common, this conclusion
 is of practical importance.
 
@@ -512,13 +512,13 @@
 \vspace*{-3mm}
 \centerline{\resizebox{.99\textwidth}{!}{\includegraphics{images/improvements_charts.pdf}}}
 \vspace*{-3mm}
-\caption{Relative improvement in error rate due to self-taught learning. 
+\caption{Relative improvement in error rate due to out-of-distribution examples.
 Left: Improvement (or loss, when negative)
 induced by out-of-distribution examples (perturbed data). 
 Right: Improvement (or loss, when negative) induced by multi-task 
 learning (training on all classes and testing only on either digits,
 upper case, or lower-case). The deep learner (SDA) benefits more from
-both self-taught learning scenarios, compared to the shallow MLP.}
+out-of-distribution examples, compared to the shallow MLP.}
 \label{fig:improvements-charts}
 \vspace*{-2mm}
 \end{figure*}
@@ -557,8 +557,9 @@
 
 In addition, as shown in the left of
 Figure~\ref{fig:improvements-charts}, the relative improvement in error
-rate brought by self-taught learning is greater for the SDA, and these
-differences with the MLP are statistically and qualitatively
+rate brought by out-of-distribution examples is greater for the deep
+stacked SDA, and these
+differences with the shallow MLP are statistically and qualitatively
 significant. 
 The left side of the figure shows the improvement to the clean
 NIST test set error brought by the use of out-of-distribution examples
@@ -623,7 +624,8 @@
 \section{Conclusions and Discussion}
 \vspace*{-2mm}
 
-We have found that the self-taught learning framework is more beneficial
+We have found that out-of-distribution examples (multi-task learning
+and perturbed examples) are more beneficial
 to a deep learner than to a traditional shallow and purely
 supervised learner. More precisely, 
 the answers are positive for all the questions asked in the introduction.
@@ -639,13 +641,13 @@
 and beating previously published results on the same data.
 
 $\bullet$ %\item 
-{\bf To what extent do self-taught learning scenarios help deep learners,
+{\bf To what extent do out-of-distribution examples help deep learners,
 and do they help them more than shallow supervised ones}?
 We found that distorted training examples not only made the resulting
 classifier better on similarly perturbed images but also on
 the {\em original clean examples}, and more importantly and more novel,
 that deep architectures benefit more from such {\em out-of-distribution}
-examples. MLPs were helped by perturbed training examples when tested on perturbed input 
+examples. Shallow MLPs were helped by perturbed training examples when tested on perturbed input 
 images (65\% relative improvement on NISTP) 
 but only marginally helped (5\% relative improvement on all classes) 
 or even hurt (10\% relative loss on digits)
@@ -669,7 +671,8 @@
 i.e., here, the relative gain of self-taught learning is probably preserved
 in the asymptotic regime.
 
-{\bf Why would deep learners benefit more from the self-taught learning framework}?
+{\bf Why would deep learners benefit more from the self-taught learning 
+framework and out-of-distribution examples}?
 The key idea is that the lower layers of the predictor compute a hierarchy
 of features that can be shared across tasks or across variants of the
 input distribution. A theoretical analysis of generalization improvements
@@ -692,14 +695,14 @@
 it was found that online learning on a huge dataset did not make the
 advantage of the deep learning bias vanish, and a similar phenomenon
 may be happening here. We hypothesize that unsupervised pre-training
-of a deep hierarchy with self-taught learning initializes the
+of a deep hierarchy with out-of-distribution examples initializes the
 model in the basin of attraction of supervised gradient descent
 that corresponds to better generalization. Furthermore, such good
 basins of attraction are not discovered by pure supervised learning
-(with or without self-taught settings) from random initialization, and more labeled examples
+(with or without out-of-distribution examples) from random initialization, and more labeled examples
 does not allow the shallow or purely supervised models to discover
 the kind of better basins associated
-with deep learning and self-taught learning.
+with deep learning and out-of-distribution examples.
  
 A Flash demo of the recognizer (where both the MLP and the SDA can be compared) 
 can be executed on-line at the anonymous site {\tt http://deep.host22.com}.