Mercurial > ift6266
changeset 537:47894d0ecbde
merge
author | Dumitru Erhan <dumitru.erhan@gmail.com> |
---|---|
date | Tue, 01 Jun 2010 18:28:43 -0700 |
parents | 5157a5830125 (diff) 22d5cd82d5f0 (current diff) |
children | f0ee2212ea7c |
files | writeup/nips2010_submission.tex |
diffstat | 1 files changed, 1 insertions(+), 1 deletions(-) [+] |
line wrap: on
line diff
--- a/writeup/nips2010_submission.tex Tue Jun 01 21:24:39 2010 -0400 +++ b/writeup/nips2010_submission.tex Tue Jun 01 18:28:43 2010 -0700 @@ -692,7 +692,7 @@ experiments showed its positive effects in a \emph{limited labeled data} scenario. However, many of the results by \citet{RainaR2007} (who used a shallow, sparse coding approach) suggest that the relative gain of self-taught -learning diminishes as the number of labeled examples increases, (essentially, +learning diminishes as the number of labeled examples increases (essentially, a ``diminishing returns'' scenario occurs). We note instead that, for deep architectures, our experiments show that such a positive effect is accomplished even in a scenario with a \emph{very large number of labeled examples}.