Mercurial > ift6266
diff writeup/contributions.tex @ 586:f5a198b2854a
contributions.tex
author | Yoshua Bengio <bengioy@iro.umontreal.ca> |
---|---|
date | Thu, 30 Sep 2010 17:43:48 -0400 |
parents | |
children |
line wrap: on
line diff
--- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/writeup/contributions.tex Thu Sep 30 17:43:48 2010 -0400 @@ -0,0 +1,97 @@ +\documentclass{article} % For LaTeX2e +\usepackage{times} +\usepackage{wrapfig} +\usepackage{amsthm,amsmath,bbm} +\usepackage[psamsfonts]{amssymb} +\usepackage{algorithm,algorithmic} +\usepackage[utf8]{inputenc} +\usepackage{graphicx,subfigure} +\usepackage[numbers]{natbib} + +\addtolength{\textwidth}{10mm} +\addtolength{\evensidemargin}{-5mm} +\addtolength{\oddsidemargin}{-5mm} + +%\setlength\parindent{0mm} + +\begin{document} + +\begin{center} +{\Large Deep Self-Taught Learning for Handwritten Character Recognition} + +{\bf \large Information on Main Contributions} +\end{center} + +\setlength{\parindent}{0cm} + +%\vspace*{-2mm} +\section*{Background and Related Contributions} +%\vspace*{-2mm} +%{\large \bf Background and Related Contributions} + +Recent theoretical and empirical work in statistical machine learning has +demonstrated the potential of learning algorithms for {\bf deep + architectures}, i.e., function classes obtained by composing multiple +levels of representation +\citep{Hinton06,ranzato-07-small,Bengio-nips-2006,VincentPLarochelleH2008,ranzato-08,Larochelle-jmlr-2009,Salakhutdinov+Hinton-2009,HonglakL2009,HonglakLNIPS2009,Jarrett-ICCV2009,Taylor-cvpr-2010}. +See~\citet{Bengio-2009} for a review of deep learning algorithms. + +{\bf Self-taught learning}~\citep{RainaR2007} is a paradigm that combines +principles of semi-supervised and multi-task learning: the learner can +exploit examples that are unlabeled and possibly come from a distribution +different from the target distribution, e.g., from other classes than those +of interest. Self-taught learning has already been applied to deep +learners, but mostly to show the advantage of unlabeled +examples~\citep{Bengio-2009,WestonJ2008-small}. + +There already are theoretical arguments~\citep{baxter95a} supporting the claim +that learning an {\bf intermediate representation} shared across tasks can be +beneficial for multi-task learning. It has also already been argued~\citep{Bengio-2009} +that {\bf multiple levels of representation} can bring a benefit over a single level. + +%{\large \bf Main Claim} +%\vspace*{-2mm} +\section*{Main Claim} +%\vspace*{-2mm} + +We claim that deep learners, with several levels of representation, can +benefit more from self-taught learning than shallow learners (with a single +level), both in the context of the multi-task setting and from {\em + out-of-distribution examples} in general. + +%{\large \bf Contribution to Machine Learning} +%\vspace*{-2mm} +\section*{Contribution to Machine Learning} +%\vspace*{-2mm} + +We show evidence for the above claim in a large-scale setting, with +a training set consisting of hundreds of millions of examples, in the +context of handwritten character recognition with 62 classes (upper-case, +lower-case, digits). + +%{\large \bf Evidence to Support the Claim} +%\vspace*{-2mm} +\section*{Evidence to Support the Claim} +%\vspace*{-2mm} + +In the above experimental setting, we show that {\em deep learners benefited +significantly more from the multi-task setting than a corresponding shallow + learner}. and that they benefited more from {\em distorted (out-of-distribution) examples} +(i.e. from a distribution larger than the one from which test examples come from). + +In addition, we show that they {\em beat previously published results} on this task +(the MNIST special database 19) +and {\bf reach human-level performance} on both handwritten digit classification and +62-class handwritten character recognition. + +\newpage + +{\small +\bibliography{strings,strings-short,strings-shorter,ift6266_ml,specials,aigaion-shorter} +%\bibliographystyle{plainnat} +\bibliographystyle{unsrtnat} +%\bibliographystyle{apalike} +} + + +\end{document}