comparison writeup/nips2010_submission.tex @ 482:ce69aa9204d8

changement au titre et reecriture abstract
author Yoshua Bengio <bengioy@iro.umontreal.ca>
date Mon, 31 May 2010 13:59:11 -0400
parents 150203d2b5c3
children b9cdb464de5f
comparison
equal deleted inserted replaced
481:3e4290448eeb 482:ce69aa9204d8
5 \usepackage{algorithm,algorithmic} 5 \usepackage{algorithm,algorithmic}
6 \usepackage[utf8]{inputenc} 6 \usepackage[utf8]{inputenc}
7 \usepackage{graphicx,subfigure} 7 \usepackage{graphicx,subfigure}
8 \usepackage[numbers]{natbib} 8 \usepackage[numbers]{natbib}
9 9
10 \title{Generating and Exploiting Perturbed and Multi-Task Handwritten Training Data for Deep Architectures} 10 \title{Deep Self-Taught Learning for Handwritten Character Recognition}
11 \author{The IFT6266 Gang} 11 \author{The IFT6266 Gang}
12 12
13 \begin{document} 13 \begin{document}
14 14
15 %\makeanontitle 15 %\makeanontitle
16 \maketitle 16 \maketitle
17 17
18 \begin{abstract} 18 \begin{abstract}
19 Recent theoretical and empirical work in statistical machine learning has 19 Recent theoretical and empirical work in statistical machine learning has
20 demonstrated the importance of learning algorithms for deep 20 demonstrated the importance of learning algorithms for deep
21 architectures, i.e., function classes obtained by composing multiple 21 architectures, i.e., function classes obtained by composing multiple
22 non-linear transformations. In the area of handwriting recognition, 22 non-linear transformations. The self-taught learning (exploitng unlabeled
23 deep learning algorithms 23 examples or examples from other distributions) has already been applied
24 had been evaluated on rather small datasets with a few tens of thousands 24 to deep learners, but mostly to show the advantage of unlabeled
25 of examples. Here we propose a powerful generator of variations 25 examples. Here we explore the advantage brought by {\em out-of-distribution
26 of examples for character images based on a pipeline of stochastic 26 examples} and show that {\em deep learners benefit more from them than a
27 transformations that include not only the usual affine transformations 27 corresponding shallow learner}, in the area
28 but also the addition of slant, local elastic deformations, changes 28 of handwritten character recognition. In fact, we show that they reach
29 in thickness, background images, color, contrast, occlusion, and 29 human-level performance on both handwritten digit classification and
30 various types of pixel and spatially correlated noise. 30 62-class handwritten character recognition. For this purpose we
31 We evaluate a deep learning algorithm (Stacked Denoising Autoencoders) 31 developed a powerful generator of stochastic variations and noise
32 on the task of learning to classify digits and letters transformed 32 processes character images, including not only affine transformations but
33 with this pipeline, using the hundreds of millions of generated examples 33 also slant, local elastic deformations, changes in thickness, background
34 and testing on the full 62-class NIST test set. 34 images, color, contrast, occlusion, and various types of pixel and
35 We find that the SDA outperforms its 35 spatially correlated noise. The out-of-distribution examples are
36 shallow counterpart, an ordinary Multi-Layer Perceptron, 36 obtained by training with these highly distorted images or
37 and that it is better able to take advantage of the additional 37 by including object classes different from those in the target test set.
38 generated data, as well as better able to take advantage of
39 the multi-task setting, i.e.,
40 training from more classes than those of interest in the end.
41 In fact, we find that the SDA reaches human performance as
42 estimated by the Amazon Mechanical Turk on the 62-class NIST test characters.
43 \end{abstract} 38 \end{abstract}
44 39
45 \section{Introduction} 40 \section{Introduction}
46 41
47 Deep Learning has emerged as a promising new area of research in 42 Deep Learning has emerged as a promising new area of research in
277 \end{figure} 272 \end{figure}
278 273
279 274
280 \begin{figure}[h] 275 \begin{figure}[h]
281 \resizebox{.99\textwidth}{!}{\includegraphics{images/transfo.png}}\\ 276 \resizebox{.99\textwidth}{!}{\includegraphics{images/transfo.png}}\\
282 \caption{Illustration of each transformation applied to the same image 277 \caption{Illustration of each transformation applied alone to the same image
283 of the upper-case h (upper-left image). first row (from left to rigth) : original image, slant, 278 of an upper-case h (top left). First row (from left to rigth) : original image, slant,
284 thickness, affine transformation, local elastic deformation; second row (from left to rigth) : 279 thickness, affine transformation, local elastic deformation; second row (from left to rigth) :
285 pinch, motion blur, occlusion, pixel permutation, gaussian noise; third row (from left to rigth) : 280 pinch, motion blur, occlusion, pixel permutation, Gaussian noise; third row (from left to rigth) :
286 background image, salt and pepper noise, spatially gaussian noise, scratches, 281 background image, salt and pepper noise, spatially Gaussian noise, scratches,
287 color and contrast changes.} 282 color and contrast changes.}
288 \label{fig:transfo} 283 \label{fig:transfo}
289 \end{figure} 284 \end{figure}
290 285
291 286