# HG changeset patch # User Xavier Glorot # Date 1243608500 14400 # Node ID 4617ee2776983945b199e9ba634b284b35bead9e # Parent 7ab4bc96cb12064b9457e5ad211ed7f917661422 fix a bad declaration of the updates in StackedDAAig diff -r 7ab4bc96cb12 -r 4617ee277698 pylearn/algorithms/sandbox/DAA_inputs_groups.py --- a/pylearn/algorithms/sandbox/DAA_inputs_groups.py Thu May 28 12:12:34 2009 -0400 +++ b/pylearn/algorithms/sandbox/DAA_inputs_groups.py Fri May 29 10:48:20 2009 -0400 @@ -529,8 +529,8 @@ self.totalgradients[-1] = [T.grad(self.totalcost[-2], paramstot) ,\ T.grad(self.globalcost[-1], paramstot) ] - local_grads = dict((j, j - self.unsup_lr * g) for j,g in zip(self.daaig[-1].params,self.localgradients[-1])) - global_grads = dict((j, j - self.unsup_lr * g)\ + local_grads = dict((j, j - self.sup_lr * g) for j,g in zip(self.daaig[-1].params,self.localgradients[-1])) + global_grads = dict((j, j - self.sup_lr * g)\ for j,g in zip(self.daaig[-1].params+paramsenc,self.globalgradients[-1])) if self.totalupdatebool: total_grads = dict((j, j - self.unsup_lr * g1 - self.sup_lr * g2)\