# HG changeset patch # User James Bergstra # Date 1210623442 14400 # Node ID cc8b032417dbeb428a6721f542fc9fa3345a351b # Parent ae5651a3696b9981b546c6ccc07b7534813c64bc# Parent f8a1ae7eb83e1ece601e590f03a1016d8c61ee6d merged diff -r f8a1ae7eb83e -r cc8b032417db mlp.py --- a/mlp.py Mon May 12 16:13:48 2008 -0400 +++ b/mlp.py Mon May 12 16:17:22 2008 -0400 @@ -89,7 +89,7 @@ self._regularization_term = self._L2_regularizer * (t.sum(self._W1*self._W1) + t.sum(self._W2*self._W2)) self._output_activations =self._b2+t.dot(t.tanh(self._b1+t.dot(self._input,self._W1.T)),self._W2.T) self._nll,self._output = crossentropy_softmax_1hot(self._output_activations,self._target_vector) - self._output_class, self._max_output = t.argmax(self._output,1) + self._output_class = t.argmax(self._output,1) self._class_error = t.neq(self._output_class,self._target_vector) self._minibatch_criterion = self._nll + self._regularization_term / t.shape(self._input)[0] OnlineGradientTLearner.__init__(self)