view doc/v2_planning/optimization.txt @ 1025:1c96e7ad95c3

coding_style: Added discussion point about backward compatibility with Python 2.x versions
author Olivier Delalleau <delallea@iro>
date Fri, 03 Sep 2010 16:58:25 -0400
parents 618b9fdbfda5
children a1b6ccd5b6dc
line wrap: on
line source

Discussion of Optimization-Related Issues
=========================================

Members: JB, PL, OD

Representative: JB


Previous work - scikits, openopt, scipy  provide function optimization
algorithms.  These are not currently GPU-enabled but may be in the future.


IS PREVIOUS WORK SUFFICIENT?
--------------------------------

In many cases it is (I used it for sparse coding, and it was ok).

These packages provide batch optimization, whereas we typically need online
optimization.

It can be faster (to run) and more convenient (to implement) to have
optimization algorithms as Theano update expressions.


What optimization algorithms do we want/need?
---------------------------------------------

 - sgd 
 - sgd + momentum
 - sgd with annealing schedule
 - TONGA
 - James Marten's Hessian-free

Do we need anything to make batch algos work better with Pylearn things?
 - conjugate methods?
 - L-BFGS?