changeset 532:2e33885730cf

changements aux charts.ods
author Yoshua Bengio <bengioy@iro.umontreal.ca>
date Tue, 01 Jun 2010 21:19:54 -0400
parents 4354c3c8f49c
children 22d5cd82d5f0
files writeup/images/charts.ods writeup/nips2010_submission.tex
diffstat 2 files changed, 7 insertions(+), 6 deletions(-) [+]
line wrap: on
line diff
Binary file writeup/images/charts.ods has changed
--- a/writeup/nips2010_submission.tex	Tue Jun 01 20:48:05 2010 -0400
+++ b/writeup/nips2010_submission.tex	Tue Jun 01 21:19:54 2010 -0400
@@ -85,12 +85,13 @@
 Self-taught learning~\citep{RainaR2007} is a paradigm that combines principles
 of semi-supervised and multi-task learning: the learner can exploit examples
 that are unlabeled and/or come from a distribution different from the target
-distribution, e.g., from other classes that those of interest. Whereas
-it has already been shown that deep learners can clearly take advantage of
-unsupervised learning and unlabeled examples~\citep{Bengio-2009,WestonJ2008-small}
-and multi-task learning, not much has been done yet to explore the impact
+distribution, e.g., from other classes that those of interest. 
+It has already been shown that deep learners can clearly take advantage of
+unsupervised learning and unlabeled examples~\citep{Bengio-2009,WestonJ2008-small},
+but more needs to be done to explore the impact
 of {\em out-of-distribution} examples and of the multi-task setting
-(but see~\citep{CollobertR2008}). In particular the {\em relative
+(one exception is~\citep{CollobertR2008}, but using very different kinds
+of learning algorithms). In particular the {\em relative
 advantage} of deep learning for this settings has not been evaluated.
 The hypothesis explored here is that a deep hierarchy of features
 may be better able to provide sharing of statistical strength
@@ -510,7 +511,7 @@
 Here we chose to use the Denoising
 Auto-Encoder~\citep{VincentPLarochelleH2008} as the building block for
 these deep hierarchies of features, as it is very simple to train and
-teach (see Figure~\ref{fig:da}, as well as 
+explain (see Figure~\ref{fig:da}, as well as 
 tutorial and code there: {\tt http://deeplearning.net/tutorial}), 
 provides immediate and efficient inference, and yielded results
 comparable or better than RBMs in series of experiments