Mercurial > pylearn
diff algorithms/tests/test_daa.py @ 500:3c60c2db0319
Added new daa test
author | Joseph Turian <turian@gmail.com> |
---|---|
date | Tue, 28 Oct 2008 13:36:27 -0400 |
parents | 32509c479e2d |
children | 4fb6f7320518 |
line wrap: on
line diff
--- a/algorithms/tests/test_daa.py Tue Oct 28 12:57:49 2008 -0400 +++ b/algorithms/tests/test_daa.py Tue Oct 28 13:36:27 2008 -0400 @@ -5,6 +5,7 @@ import numpy import time +import pylearn.algorithms.logistic_regression def test_train_daa(mode = theano.Mode('c|py', 'fast_run')): @@ -30,6 +31,30 @@ print model.classify([[0, 1, 0, 1]]) +def test_train_daa2(mode = theano.Mode('c|py', 'fast_run')): + + ndaa = 3 + daa = models.Stacker([(models.SigmoidXEDenoisingAA, 'hidden')] * ndaa + [(pylearn.algorithms.logistic_regression.Module_Nclass, 'output')], + regularize = False) + + model = daa.make([4, 20, 20, 20, 1], + lr = 0.01, + mode = mode, + seed = 10) + + model.layers[0].noise_level = 0.3 + model.layers[1].noise_level = 0.3 + model.layers[2].noise_level = 0.3 + + # Update the first hidden layer + model.local_update[0]([[0, 1, 0, 1]]) + model.local_update[1]([[0, 1, 0, 1]]) + model.local_update[2]([[0, 1, 0, 1]]) + + model.update([[0, 1, 0, 1]], [0]) + print model.classify([[0, 1, 0, 1]]) + + @@ -46,3 +71,4 @@ # test_train_daa(theano.compile.Mode('c&py', 'merge')) test_train_daa(theano.compile.Mode('c|py', 'merge')) + test_train_daa2(theano.compile.Mode('c|py', 'merge'))