Mercurial > pylearn
diff sandbox/simple_autoassociator/model.py @ 417:4f61201fa9a9
Parameters are no longer global
author | Joseph Turian <turian@iro.umontreal.ca> |
---|---|
date | Fri, 11 Jul 2008 17:19:37 -0400 |
parents | 8849eba55520 |
children |
line wrap: on
line diff
--- a/sandbox/simple_autoassociator/model.py Fri Jul 11 16:34:46 2008 -0400 +++ b/sandbox/simple_autoassociator/model.py Fri Jul 11 17:19:37 2008 -0400 @@ -6,22 +6,30 @@ from graph import trainfn import parameters -import globals -from globals import LR - import numpy import random -random.seed(globals.SEED) import pylearn.sparse_instance class Model: - def __init__(self): - self.parameters = parameters.Parameters(randomly_initialize=True) + """ + @todo: Add momentum. + @todo: Add learning rate decay schedule. + """ + def __init__(self, input_dimension, hidden_dimension, learning_rate = 0.1, weight_decay = 0.0002, random_seed = 666): + self.input_dimension = input_dimension + self.hidden_dimension = hidden_dimension + self.learning_rate = learning_rate + self.weight_decay = weight_decay + self.random_seed = random_seed -# def deterministic_reconstruction(self, x): -# (y, h, loss, gw1, gb1, gw2, gb2) = trainfn(x, self.parameters.w1, self.parameters.b1, self.parameters.w2, self.parameters.b2) -# return y + random.seed(random_seed) + + self.parameters = parameters.Parameters(input_dimension=self.input_dimension, hidden_dimension=self.hidden_dimension, randomly_initialize=True, random_seed=self.random_seed) + + def deterministic_reconstruction(self, x): + (y, h, loss, gw1, gb1, gw2, gb2) = trainfn(x, self.parameters.w1, self.parameters.b1, self.parameters.w2, self.parameters.b2) + return y def update(self, instances): """ @@ -29,10 +37,11 @@ @param instances: A list of dict from feature index to (non-zero) value. @todo: Should assert that nonzero_indices and zero_indices are correct (i.e. are truly nonzero/zero). + @todo: Multiply L{self.weight_decay} by L{self.learning_rate}, as done in Semantic Hashing? + @todo: Decay the biases too? """ minibatch = len(instances) -# x = pylearn.sparse_instance.to_vector(instances, self.input_dimension) - x = pylearn.sparse_instance.to_vector(instances, globals.INPUT_DIMENSION) + x = pylearn.sparse_instance.to_vector(instances, self.input_dimension) (y, h, loss, gw1, gb1, gw2, gb2) = trainfn(x, self.parameters.w1, self.parameters.b1, self.parameters.w2, self.parameters.b2) # print @@ -45,15 +54,18 @@ # print "gw2:", gw2 # print "gb2:", gb2 - # SGD update - self.parameters.w1 -= LR * gw1 - self.parameters.b1 -= LR * gb1 - self.parameters.w2 -= LR * gw2 - self.parameters.b2 -= LR * gb2 + self.parameters.w1 *= (1 - self.weight_decay) + self.parameters.w2 *= (1 - self.weight_decay) - # Recompute the loss, to make sure it's descreasing - (y, h, loss, gw1, gb1, gw2, gb2) = trainfn(x, self.parameters.w1, self.parameters.b1, self.parameters.w2, self.parameters.b2) -# print "NEW y:", y - print "NEW total loss:", loss -# print "h:", h -# print self.parameters + # SGD update + self.parameters.w1 -= self.learning_rate * gw1 / minibatch + self.parameters.b1 -= self.learning_rate * gb1 / minibatch + self.parameters.w2 -= self.learning_rate * gw2 / minibatch + self.parameters.b2 -= self.learning_rate * gb2 / minibatch + +# # Recompute the loss, to make sure it's descreasing +# (y, h, loss, gw1, gb1, gw2, gb2) = trainfn(x, self.parameters.w1, self.parameters.b1, self.parameters.w2, self.parameters.b2) +## print "NEW y:", y +# print "NEW total loss:", loss +## print "h:", h +## print self.parameters