changeset 1001:660d784d14c7

moved planning into its directory
author Yoshua Bengio <bengioy@iro.umontreal.ca>
date Wed, 01 Sep 2010 12:18:08 -0400
parents d4a14c6c36e0
children f82093bf4405
files doc/v2_planning.txt doc/v2_planning/main_plan.txt
diffstat 2 files changed, 241 insertions(+), 241 deletions(-) [+]
line wrap: on
line diff
--- a/doc/v2_planning.txt	Tue Aug 24 19:24:54 2010 -0400
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,241 +0,0 @@
-
-Motivation
-==========
-
-Yoshua:
--------
-
-We are missing a *Theano Machine Learning library*.
-
-The deep learning tutorials do a good job but they lack the following features, which I would like to see in a ML library:
-
- - a well-organized collection of Theano symbolic expressions (formulas) for handling most of
-   what is needed either in implementing existing well-known ML and deep learning algorithms or
-   for creating new variants (without having to start from scratch each time), that is the
-   mathematical core,
-
- - a well-organized collection of python modules to help with the following:
-      - several data-access models that wrap around learning algorithms for interfacing with various types of data (static vectors, images, sound, video, generic time-series, etc.)
-      - generic utility code for optimization
-             - stochastic gradient descent variants
-             - early stopping variants
-             - interfacing to generic 2nd order optimization methods
-             - 2nd order methods tailored to work on minibatches
-             - optimizers for sparse coefficients / parameters
-     - generic code for model selection and hyper-parameter optimization (including the use and coordination of multiple jobs running on different machines, e.g. using jobman)
-     - generic code for performance estimation and experimental statistics
-     - visualization tools (using existing python libraries) and examples for all of the above
-     - learning algorithm conventions and meta-learning algorithms (bagging, boosting, mixtures of experts, etc.) which use them
-
-   [Note that many of us already use some instance of all the above, but each one tends to reinvent the wheel and newbies don't benefit from a knowledge base.]
-
- - a well-documented set of python scripts using the above library to show how to run the most
-   common ML algorithms (possibly with examples showing how to run multiple experiments with
-   many different models and collect statistical comparative results). This is particularly
-   important for pure users to adopt Theano in the ML application work.
-
-Ideally, there would be one person in charge of this project, making sure a coherent and
-easy-to-read design is developed, along with many helping hands (to implement the various
-helper modules, formulae, and learning algorithms).
-
-
-James:
--------
-
-I am interested in the design and implementation of the "well-organized collection of Theano
-symbolic expressions..."
-
-I would like to explore algorithms for hyper-parameter optimization, following up on some
-"high-throughput" work.  I'm most interested in the "generic code for model selection and
-hyper-parameter optimization..." and "generic code for performance estimation...".  
-
-I have some experiences with the data-access requirements, and some lessons I'd like to share
-on that, but no time to work on that aspect of things.
-
-I will continue to contribute to the "well-documented set of python scripts using the above to
-showcase common ML algorithms...".  I have an Olshausen&Field-style sparse coding script that
-could be polished up.  I am also implementing the mcRBM and I'll be able to add that when it's
-done.
-
-
-
-Suggestions for how to tackle various desiderata
-================================================
-
-
-Theano Symbolic Expressions for ML
-----------------------------------
-
-We could make this a submodule of pylearn: ``pylearn.nnet``.  
-
-Yoshua: I would use a different name, e.g., "pylearn.formulas" to emphasize that it is not just 
-about neural nets, and that this is a collection of formulas (expressions), rather than
-completely self-contained classes for learners. We could have a "nnet.py" file for
-neural nets, though.
-
-There are a number of ideas floating around for how to handle classes /
-modules (LeDeepNet, pylearn.shared.layers, pynnet, DeepAnn) so lets implement as much
-math as possible in global functions with no classes.  There are no models in
-the wish list that require than a few vectors and matrices to parametrize.
-Global functions are more reusable than classes.
-
-
-Data access 
------------
-
-A general interface to datasets from the perspective of an experiment driver
-(e.g. kfold) is to see them as a function that maps index (typically integer)
-to example (whose type and nature depends on the dataset, it could for
-instance be an (image, label) pair).  This interface permits iterating over
-the dataset, shuffling the dataset, and splitting it into folds.  For
-efficiency, it is nice if the dataset interface supports looking up several
-index values at once, because looking up many examples at once can sometimes
-be faster than looking each one up in turn. In particular, looking up
-a consecutive block of indices, or a slice, should be well supported.
-
-Some datasets may not support random access (e.g. a random number stream) and
-that's fine if an exception is raised. The user will see a NotImplementedError
-or similar, and try something else. We might want to have a way to test
-that a dataset is random-access or not without having to load an example.
-
-
-A more intuitive interface for many datasets (or subsets) is to load them as
-matrices or lists of examples.  This format is more convenient to work with at
-an ipython shell, for example.  It is not good to provide only the "dataset
-as a function" view of a dataset.  Even if a dataset is very large, it is nice
-to have a standard way to get some representative examples in a convenient
-structure, to be able to play with them in ipython.
-
-
-Another thing to consider related to datasets is that there are a number of
-other efforts to have standard ML datasets, and we should be aware of them,
-and compatible with them when it's easy:
- - mldata.org    (they have a file format, not sure how many use it)
- - weka          (ARFF file format)
- - scikits.learn 
- - hdf5 / pytables
-
-
-pylearn.datasets uses a DATA_ROOT environment variable to locate a filesystem
-folder that is assumed to have a standard form across different installations.
-That's where the data files are.  The correct format of this folder is currently
-defined implicitly by the contents of /data/lisa/data at DIRO, but it would be
-better to document in pylearn what the contents of this folder should be as
-much as possible.  It should be possible to rebuild this tree from information
-found in pylearn.
-
-Yoshua (about ideas proposed by Pascal Vincent a while ago): 
-
-  - we may want to distinguish between datasets and tasks: a task defines
-  not just the data but also things like what is the input and what is the
-  target (for supervised learning), and *importantly* a set of performance metrics
-  that make sense for this task (e.g. those used by papers solving a particular
-  task, or reported for a particular benchmark)
-
-  - we should discuss about a few "standards" that datasets and tasks may comply to, such as
-    - "input" and "target" fields inside each example, for supervised or semi-supervised learning tasks
-      (with a convention for the semi-supervised case when only the input or only the target is observed)
-    - "input" for unsupervised learning
-    - conventions for missing-valued components inside input or target 
-    - how examples that are sequences are treated (e.g. the input or the target is a sequence)
-    - how time-stamps are specified when appropriate (e.g., the sequences are asynchronous)
-    - how error metrics are specified
-        * example-level statistics (e.g. classification error)
-        * dataset-level statistics (e.g. ROC curve, mean and standard error of error)
-
-
-Model Selection & Hyper-Parameter Optimization
-----------------------------------------------
-
-Driving a distributed computing job for a long time to optimize
-hyper-parameters using one or more clusters is the goal here.
-Although there might be some library-type code to write here, I think of this
-more as an application template.  The user would use python code to describe
-the experiment to run and the hyper-parameter space to search.  Then this
-application-driver would take control of scheduling jobs and running them on
-various computers... I'm imagining a potentially ugly brute of a hack that's
-not necessarily something we will want to expose at a low-level for reuse.
-
-Yoshua: We want both the library-defined driver that takes instructions about how to generate
-new hyper-parameter combinations (e.g. implicitly providing a prior distribution from which 
-to sample them), and examples showing how to use it in typical cases.
-Note that sometimes we just want to find the best configuration of hyper-parameters,
-but sometimes we want to do more subtle analysis. Often a combination of both.
-In this respect it could be useful for the user to define hyper-parameters over
-which scientific questions are sought (e.g. depth of an architecture) vs
-hyper-parameters that we would like to marginalize/maximize over (e.g. learning rate).
-This can influence both the sampling of configurations (we want to make sure that all
-combinations of question-driving hyper-parameters are covered) and the analysis
-of results (we may be willing to estimate ANOVAs or averaging or quantiles over
-the non-question-driving hyper-parameters).
-
-Python scripts for common ML algorithms
----------------------------------------
-
-The script aspect of this feature request makes me think that what would be
-good here is more tutorial-type scripts.  And the existing tutorials could
-potentially be rewritten to use some of the pylearn.nnet expressions.   More
-tutorials / demos would be great.
-
-Yoshua: agreed that we could write them as tutorials, but note how the
-spirit would be different from the current deep learning tutorials: we would
-not mind using library code as much as possible instead of trying to flatten
-out everything in the interest of pedagogical simplicity. Instead, these
-tutorials should be meant to illustrate not the algorithms but *how to take
-advantage of the library*. They could also be used as *BLACK BOX* implementations
-by people who don't want to dig lower and just want to run experiments.
-
-Functional Specifications
-=========================
-
-TODO: 
-Put these into different text files so that this one does not become a monster.
-For each thing with a functional spec (e.g. datasets library, optimization library) make a
-separate file.
-
-
-
-pylearn.formulas
-----------------
-
-Directory with functions for building layers, calculating classification
-errors, cross-entropies with various distributions, free energies, etc.  This
-module would include for the most part global functions, Theano Ops and Theano
-optimizations.
-
-Yoshua: I would break it down in module files, e.g.:
-
-pylearn.formulas.costs: generic / common cost functions, e.g. various cross-entropies, squared error, 
-abs. error, various sparsity penalties (L1, Student)
-
-pylearn.formulas.linear: formulas for linear classifier, linear regression, factor analysis, PCA
-
-pylearn.formulas.nnet: formulas for building layers of various kinds, various activation functions,
-layers which could be plugged with various costs & penalties, and stacked
-
-pylearn.formulas.ae: formulas for auto-encoders and denoising auto-encoder variants
-
-pylearn.formulas.noise: formulas for corruption processes
-
-pylearn.formulas.rbm: energies, free energies, conditional distributions, Gibbs sampling
-
-pylearn.formulas.trees: formulas for decision trees
-
-pylearn.formulas.boosting: formulas for boosting variants
-
-etc.
-
-Fred: It seam that the DeepANN git repository by Xavier G. have part of this as function.
-
-Indexing Convention
-~~~~~~~~~~~~~~~~~~~
-
-Something to decide on - Fortran-style or C-style indexing.  Although we have
-often used c-style indexing in the past (for efficiency in c!) this is no
-longer an issue with numpy because the physical layout is independent of the
-indexing order.  The fact remains that Fortran-style indexing follows linear
-algebra conventions, while c-style indexing does not.  If a global function
-includes a lot of math derivations, it would be *really* nice if the code used
-the same convention for the orientation of matrices, and endlessly annoying to
-have to be always transposing everything.
-
--- /dev/null	Thu Jan 01 00:00:00 1970 +0000
+++ b/doc/v2_planning/main_plan.txt	Wed Sep 01 12:18:08 2010 -0400
@@ -0,0 +1,241 @@
+
+Motivation
+==========
+
+Yoshua:
+-------
+
+We are missing a *Theano Machine Learning library*.
+
+The deep learning tutorials do a good job but they lack the following features, which I would like to see in a ML library:
+
+ - a well-organized collection of Theano symbolic expressions (formulas) for handling most of
+   what is needed either in implementing existing well-known ML and deep learning algorithms or
+   for creating new variants (without having to start from scratch each time), that is the
+   mathematical core,
+
+ - a well-organized collection of python modules to help with the following:
+      - several data-access models that wrap around learning algorithms for interfacing with various types of data (static vectors, images, sound, video, generic time-series, etc.)
+      - generic utility code for optimization
+             - stochastic gradient descent variants
+             - early stopping variants
+             - interfacing to generic 2nd order optimization methods
+             - 2nd order methods tailored to work on minibatches
+             - optimizers for sparse coefficients / parameters
+     - generic code for model selection and hyper-parameter optimization (including the use and coordination of multiple jobs running on different machines, e.g. using jobman)
+     - generic code for performance estimation and experimental statistics
+     - visualization tools (using existing python libraries) and examples for all of the above
+     - learning algorithm conventions and meta-learning algorithms (bagging, boosting, mixtures of experts, etc.) which use them
+
+   [Note that many of us already use some instance of all the above, but each one tends to reinvent the wheel and newbies don't benefit from a knowledge base.]
+
+ - a well-documented set of python scripts using the above library to show how to run the most
+   common ML algorithms (possibly with examples showing how to run multiple experiments with
+   many different models and collect statistical comparative results). This is particularly
+   important for pure users to adopt Theano in the ML application work.
+
+Ideally, there would be one person in charge of this project, making sure a coherent and
+easy-to-read design is developed, along with many helping hands (to implement the various
+helper modules, formulae, and learning algorithms).
+
+
+James:
+-------
+
+I am interested in the design and implementation of the "well-organized collection of Theano
+symbolic expressions..."
+
+I would like to explore algorithms for hyper-parameter optimization, following up on some
+"high-throughput" work.  I'm most interested in the "generic code for model selection and
+hyper-parameter optimization..." and "generic code for performance estimation...".  
+
+I have some experiences with the data-access requirements, and some lessons I'd like to share
+on that, but no time to work on that aspect of things.
+
+I will continue to contribute to the "well-documented set of python scripts using the above to
+showcase common ML algorithms...".  I have an Olshausen&Field-style sparse coding script that
+could be polished up.  I am also implementing the mcRBM and I'll be able to add that when it's
+done.
+
+
+
+Suggestions for how to tackle various desiderata
+================================================
+
+
+Theano Symbolic Expressions for ML
+----------------------------------
+
+We could make this a submodule of pylearn: ``pylearn.nnet``.  
+
+Yoshua: I would use a different name, e.g., "pylearn.formulas" to emphasize that it is not just 
+about neural nets, and that this is a collection of formulas (expressions), rather than
+completely self-contained classes for learners. We could have a "nnet.py" file for
+neural nets, though.
+
+There are a number of ideas floating around for how to handle classes /
+modules (LeDeepNet, pylearn.shared.layers, pynnet, DeepAnn) so lets implement as much
+math as possible in global functions with no classes.  There are no models in
+the wish list that require than a few vectors and matrices to parametrize.
+Global functions are more reusable than classes.
+
+
+Data access 
+-----------
+
+A general interface to datasets from the perspective of an experiment driver
+(e.g. kfold) is to see them as a function that maps index (typically integer)
+to example (whose type and nature depends on the dataset, it could for
+instance be an (image, label) pair).  This interface permits iterating over
+the dataset, shuffling the dataset, and splitting it into folds.  For
+efficiency, it is nice if the dataset interface supports looking up several
+index values at once, because looking up many examples at once can sometimes
+be faster than looking each one up in turn. In particular, looking up
+a consecutive block of indices, or a slice, should be well supported.
+
+Some datasets may not support random access (e.g. a random number stream) and
+that's fine if an exception is raised. The user will see a NotImplementedError
+or similar, and try something else. We might want to have a way to test
+that a dataset is random-access or not without having to load an example.
+
+
+A more intuitive interface for many datasets (or subsets) is to load them as
+matrices or lists of examples.  This format is more convenient to work with at
+an ipython shell, for example.  It is not good to provide only the "dataset
+as a function" view of a dataset.  Even if a dataset is very large, it is nice
+to have a standard way to get some representative examples in a convenient
+structure, to be able to play with them in ipython.
+
+
+Another thing to consider related to datasets is that there are a number of
+other efforts to have standard ML datasets, and we should be aware of them,
+and compatible with them when it's easy:
+ - mldata.org    (they have a file format, not sure how many use it)
+ - weka          (ARFF file format)
+ - scikits.learn 
+ - hdf5 / pytables
+
+
+pylearn.datasets uses a DATA_ROOT environment variable to locate a filesystem
+folder that is assumed to have a standard form across different installations.
+That's where the data files are.  The correct format of this folder is currently
+defined implicitly by the contents of /data/lisa/data at DIRO, but it would be
+better to document in pylearn what the contents of this folder should be as
+much as possible.  It should be possible to rebuild this tree from information
+found in pylearn.
+
+Yoshua (about ideas proposed by Pascal Vincent a while ago): 
+
+  - we may want to distinguish between datasets and tasks: a task defines
+  not just the data but also things like what is the input and what is the
+  target (for supervised learning), and *importantly* a set of performance metrics
+  that make sense for this task (e.g. those used by papers solving a particular
+  task, or reported for a particular benchmark)
+
+  - we should discuss about a few "standards" that datasets and tasks may comply to, such as
+    - "input" and "target" fields inside each example, for supervised or semi-supervised learning tasks
+      (with a convention for the semi-supervised case when only the input or only the target is observed)
+    - "input" for unsupervised learning
+    - conventions for missing-valued components inside input or target 
+    - how examples that are sequences are treated (e.g. the input or the target is a sequence)
+    - how time-stamps are specified when appropriate (e.g., the sequences are asynchronous)
+    - how error metrics are specified
+        * example-level statistics (e.g. classification error)
+        * dataset-level statistics (e.g. ROC curve, mean and standard error of error)
+
+
+Model Selection & Hyper-Parameter Optimization
+----------------------------------------------
+
+Driving a distributed computing job for a long time to optimize
+hyper-parameters using one or more clusters is the goal here.
+Although there might be some library-type code to write here, I think of this
+more as an application template.  The user would use python code to describe
+the experiment to run and the hyper-parameter space to search.  Then this
+application-driver would take control of scheduling jobs and running them on
+various computers... I'm imagining a potentially ugly brute of a hack that's
+not necessarily something we will want to expose at a low-level for reuse.
+
+Yoshua: We want both the library-defined driver that takes instructions about how to generate
+new hyper-parameter combinations (e.g. implicitly providing a prior distribution from which 
+to sample them), and examples showing how to use it in typical cases.
+Note that sometimes we just want to find the best configuration of hyper-parameters,
+but sometimes we want to do more subtle analysis. Often a combination of both.
+In this respect it could be useful for the user to define hyper-parameters over
+which scientific questions are sought (e.g. depth of an architecture) vs
+hyper-parameters that we would like to marginalize/maximize over (e.g. learning rate).
+This can influence both the sampling of configurations (we want to make sure that all
+combinations of question-driving hyper-parameters are covered) and the analysis
+of results (we may be willing to estimate ANOVAs or averaging or quantiles over
+the non-question-driving hyper-parameters).
+
+Python scripts for common ML algorithms
+---------------------------------------
+
+The script aspect of this feature request makes me think that what would be
+good here is more tutorial-type scripts.  And the existing tutorials could
+potentially be rewritten to use some of the pylearn.nnet expressions.   More
+tutorials / demos would be great.
+
+Yoshua: agreed that we could write them as tutorials, but note how the
+spirit would be different from the current deep learning tutorials: we would
+not mind using library code as much as possible instead of trying to flatten
+out everything in the interest of pedagogical simplicity. Instead, these
+tutorials should be meant to illustrate not the algorithms but *how to take
+advantage of the library*. They could also be used as *BLACK BOX* implementations
+by people who don't want to dig lower and just want to run experiments.
+
+Functional Specifications
+=========================
+
+TODO: 
+Put these into different text files so that this one does not become a monster.
+For each thing with a functional spec (e.g. datasets library, optimization library) make a
+separate file.
+
+
+
+pylearn.formulas
+----------------
+
+Directory with functions for building layers, calculating classification
+errors, cross-entropies with various distributions, free energies, etc.  This
+module would include for the most part global functions, Theano Ops and Theano
+optimizations.
+
+Yoshua: I would break it down in module files, e.g.:
+
+pylearn.formulas.costs: generic / common cost functions, e.g. various cross-entropies, squared error, 
+abs. error, various sparsity penalties (L1, Student)
+
+pylearn.formulas.linear: formulas for linear classifier, linear regression, factor analysis, PCA
+
+pylearn.formulas.nnet: formulas for building layers of various kinds, various activation functions,
+layers which could be plugged with various costs & penalties, and stacked
+
+pylearn.formulas.ae: formulas for auto-encoders and denoising auto-encoder variants
+
+pylearn.formulas.noise: formulas for corruption processes
+
+pylearn.formulas.rbm: energies, free energies, conditional distributions, Gibbs sampling
+
+pylearn.formulas.trees: formulas for decision trees
+
+pylearn.formulas.boosting: formulas for boosting variants
+
+etc.
+
+Fred: It seam that the DeepANN git repository by Xavier G. have part of this as function.
+
+Indexing Convention
+~~~~~~~~~~~~~~~~~~~
+
+Something to decide on - Fortran-style or C-style indexing.  Although we have
+often used c-style indexing in the past (for efficiency in c!) this is no
+longer an issue with numpy because the physical layout is independent of the
+indexing order.  The fact remains that Fortran-style indexing follows linear
+algebra conventions, while c-style indexing does not.  If a global function
+includes a lot of math derivations, it would be *really* nice if the code used
+the same convention for the orientation of matrices, and endlessly annoying to
+have to be always transposing everything.
+