Mercurial > ift6266
annotate baseline/mlp/mlp_nist.py @ 309:60cacb9a70e4
Petits changements pour pouvoir utiliser le GPU
author | SylvainPL <sylvain.pannetier.lebeuf@umontreal.ca> |
---|---|
date | Thu, 01 Apr 2010 13:43:43 -0400 |
parents | 1e4bf5a5b46d |
children | 743907366476 |
rev | line source |
---|---|
110 | 1 """ |
2 This tutorial introduces the multilayer perceptron using Theano. | |
3 | |
4 A multilayer perceptron is a logistic regressor where | |
5 instead of feeding the input to the logistic regression you insert a | |
6 intermidiate layer, called the hidden layer, that has a nonlinear | |
7 activation function (usually tanh or sigmoid) . One can use many such | |
8 hidden layers making the architecture deep. The tutorial will also tackle | |
9 the problem of MNIST digit classification. | |
10 | |
11 .. math:: | |
12 | |
13 f(x) = G( b^{(2)} + W^{(2)}( s( b^{(1)} + W^{(1)} x))), | |
14 | |
15 References: | |
16 | |
17 - textbooks: "Pattern Recognition and Machine Learning" - | |
18 Christopher M. Bishop, section 5 | |
19 | |
20 TODO: recommended preprocessing, lr ranges, regularization ranges (explain | |
21 to do lr first, then add regularization) | |
22 | |
23 """ | |
24 __docformat__ = 'restructedtext en' | |
25 | |
26 import pdb | |
27 import numpy | |
28 import pylab | |
29 import theano | |
30 import theano.tensor as T | |
31 import time | |
32 import theano.tensor.nnet | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
33 import pylearn |
304
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
34 import theano,pylearn.version,ift6266 |
110 | 35 from pylearn.io import filetensor as ft |
36 | |
37 data_path = '/data/lisa/data/nist/by_class/' | |
38 | |
39 class MLP(object): | |
40 """Multi-Layer Perceptron Class | |
41 | |
42 A multilayer perceptron is a feedforward artificial neural network model | |
43 that has one layer or more of hidden units and nonlinear activations. | |
44 Intermidiate layers usually have as activation function thanh or the | |
45 sigmoid function while the top layer is a softamx layer. | |
46 """ | |
47 | |
48 | |
49 | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
50 def __init__(self, input, n_in, n_hidden, n_out,learning_rate): |
110 | 51 """Initialize the parameters for the multilayer perceptron |
52 | |
53 :param input: symbolic variable that describes the input of the | |
54 architecture (one minibatch) | |
55 | |
56 :param n_in: number of input units, the dimension of the space in | |
57 which the datapoints lie | |
58 | |
59 :param n_hidden: number of hidden units | |
60 | |
61 :param n_out: number of output units, the dimension of the space in | |
62 which the labels lie | |
63 | |
64 """ | |
65 | |
66 # initialize the parameters theta = (W1,b1,W2,b2) ; note that this | |
67 # example contains only one hidden layer, but one can have as many | |
68 # layers as he/she wishes, making the network deeper. The only | |
69 # problem making the network deep this way is during learning, | |
70 # backpropagation being unable to move the network from the starting | |
71 # point towards; this is where pre-training helps, giving a good | |
72 # starting point for backpropagation, but more about this in the | |
73 # other tutorials | |
74 | |
75 # `W1` is initialized with `W1_values` which is uniformely sampled | |
76 # from -6./sqrt(n_in+n_hidden) and 6./sqrt(n_in+n_hidden) | |
77 # the output of uniform if converted using asarray to dtype | |
78 # theano.config.floatX so that the code is runable on GPU | |
79 W1_values = numpy.asarray( numpy.random.uniform( \ | |
80 low = -numpy.sqrt(6./(n_in+n_hidden)), \ | |
81 high = numpy.sqrt(6./(n_in+n_hidden)), \ | |
82 size = (n_in, n_hidden)), dtype = theano.config.floatX) | |
83 # `W2` is initialized with `W2_values` which is uniformely sampled | |
84 # from -6./sqrt(n_hidden+n_out) and 6./sqrt(n_hidden+n_out) | |
85 # the output of uniform if converted using asarray to dtype | |
86 # theano.config.floatX so that the code is runable on GPU | |
87 W2_values = numpy.asarray( numpy.random.uniform( | |
88 low = -numpy.sqrt(6./(n_hidden+n_out)), \ | |
89 high= numpy.sqrt(6./(n_hidden+n_out)),\ | |
90 size= (n_hidden, n_out)), dtype = theano.config.floatX) | |
91 | |
92 self.W1 = theano.shared( value = W1_values ) | |
93 self.b1 = theano.shared( value = numpy.zeros((n_hidden,), | |
94 dtype= theano.config.floatX)) | |
95 self.W2 = theano.shared( value = W2_values ) | |
96 self.b2 = theano.shared( value = numpy.zeros((n_out,), | |
97 dtype= theano.config.floatX)) | |
98 | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
99 #include the learning rate in the classifer so |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
100 #we can modify it on the fly when we want |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
101 lr_value=learning_rate |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
102 self.lr=theano.shared(value=lr_value) |
110 | 103 # symbolic expression computing the values of the hidden layer |
104 self.hidden = T.tanh(T.dot(input, self.W1)+ self.b1) | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
105 |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
106 |
110 | 107 |
108 # symbolic expression computing the values of the top layer | |
109 self.p_y_given_x= T.nnet.softmax(T.dot(self.hidden, self.W2)+self.b2) | |
110 | |
111 # compute prediction as class whose probability is maximal in | |
112 # symbolic form | |
113 self.y_pred = T.argmax( self.p_y_given_x, axis =1) | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
114 self.y_pred_num = T.argmax( self.p_y_given_x[0:9], axis =1) |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
115 |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
116 |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
117 |
110 | 118 |
119 # L1 norm ; one regularization option is to enforce L1 norm to | |
120 # be small | |
121 self.L1 = abs(self.W1).sum() + abs(self.W2).sum() | |
122 | |
123 # square of L2 norm ; one regularization option is to enforce | |
124 # square of L2 norm to be small | |
125 self.L2_sqr = (self.W1**2).sum() + (self.W2**2).sum() | |
126 | |
127 | |
128 | |
129 def negative_log_likelihood(self, y): | |
130 """Return the mean of the negative log-likelihood of the prediction | |
131 of this model under a given target distribution. | |
132 | |
133 .. math:: | |
134 | |
135 \frac{1}{|\mathcal{D}|}\mathcal{L} (\theta=\{W,b\}, \mathcal{D}) = | |
136 \frac{1}{|\mathcal{D}|}\sum_{i=0}^{|\mathcal{D}|} \log(P(Y=y^{(i)}|x^{(i)}, W,b)) \\ | |
137 \ell (\theta=\{W,b\}, \mathcal{D}) | |
138 | |
139 | |
140 :param y: corresponds to a vector that gives for each example the | |
141 :correct label | |
142 """ | |
143 return -T.mean(T.log(self.p_y_given_x)[T.arange(y.shape[0]),y]) | |
144 | |
145 | |
146 | |
147 | |
148 def errors(self, y): | |
149 """Return a float representing the number of errors in the minibatch | |
150 over the total number of examples of the minibatch | |
151 """ | |
152 | |
153 # check if y has same dimension of y_pred | |
154 if y.ndim != self.y_pred.ndim: | |
155 raise TypeError('y should have the same shape as self.y_pred', | |
156 ('y', target.type, 'y_pred', self.y_pred.type)) | |
157 # check if y is of the correct datatype | |
158 if y.dtype.startswith('int'): | |
159 # the T.neq operator returns a vector of 0s and 1s, where 1 | |
160 # represents a mistake in prediction | |
161 return T.mean(T.neq(self.y_pred, y)) | |
162 else: | |
163 raise NotImplementedError() | |
164 | |
165 | |
304
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
166 def mlp_full_nist( verbose = 1,\ |
145
8ceaaf812891
changed adaptive lr flag from bool to int for jobman issues
XavierMuller
parents:
143
diff
changeset
|
167 adaptive_lr = 0,\ |
110 | 168 train_data = 'all/all_train_data.ft',\ |
169 train_labels = 'all/all_train_labels.ft',\ | |
170 test_data = 'all/all_test_data.ft',\ | |
171 test_labels = 'all/all_test_labels.ft',\ | |
172 learning_rate=0.01,\ | |
173 L1_reg = 0.00,\ | |
174 L2_reg = 0.0001,\ | |
175 nb_max_exemples=1000000,\ | |
176 batch_size=20,\ | |
177 nb_hidden = 500,\ | |
212
e390b0454515
added classic lr time decay and py code to calculate the error based on a saved model
xaviermuller
parents:
169
diff
changeset
|
178 nb_targets = 62, |
304
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
179 tau=1e6,\ |
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
180 lr_t2_factor=0.5): |
110 | 181 |
182 | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
183 configuration = [learning_rate,nb_max_exemples,nb_hidden,adaptive_lr] |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
184 |
212
e390b0454515
added classic lr time decay and py code to calculate the error based on a saved model
xaviermuller
parents:
169
diff
changeset
|
185 #save initial learning rate if classical adaptive lr is used |
e390b0454515
added classic lr time decay and py code to calculate the error based on a saved model
xaviermuller
parents:
169
diff
changeset
|
186 initial_lr=learning_rate |
e390b0454515
added classic lr time decay and py code to calculate the error based on a saved model
xaviermuller
parents:
169
diff
changeset
|
187 |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
188 total_validation_error_list = [] |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
189 total_train_error_list = [] |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
190 learning_rate_list=[] |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
191 best_training_error=float('inf'); |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
192 |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
193 |
212
e390b0454515
added classic lr time decay and py code to calculate the error based on a saved model
xaviermuller
parents:
169
diff
changeset
|
194 |
110 | 195 |
196 f = open(data_path+train_data) | |
197 g= open(data_path+train_labels) | |
198 h = open(data_path+test_data) | |
199 i= open(data_path+test_labels) | |
200 | |
201 raw_train_data = ft.read(f) | |
202 raw_train_labels = ft.read(g) | |
203 raw_test_data = ft.read(h) | |
204 raw_test_labels = ft.read(i) | |
205 | |
206 f.close() | |
207 g.close() | |
208 i.close() | |
209 h.close() | |
210 #create a validation set the same size as the test size | |
211 #use the end of the training array for this purpose | |
212 #discard the last remaining so we get a %batch_size number | |
213 test_size=len(raw_test_labels) | |
214 test_size = int(test_size/batch_size) | |
215 test_size*=batch_size | |
216 train_size = len(raw_train_data) | |
217 train_size = int(train_size/batch_size) | |
218 train_size*=batch_size | |
219 validation_size =test_size | |
220 offset = train_size-test_size | |
304
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
221 if verbose == 1: |
110 | 222 print 'train size = %d' %train_size |
223 print 'test size = %d' %test_size | |
224 print 'valid size = %d' %validation_size | |
225 print 'offset = %d' %offset | |
226 | |
227 | |
228 train_set = (raw_train_data,raw_train_labels) | |
229 train_batches = [] | |
230 for i in xrange(0, train_size-test_size, batch_size): | |
231 train_batches = train_batches + \ | |
232 [(raw_train_data[i:i+batch_size], raw_train_labels[i:i+batch_size])] | |
233 | |
234 test_batches = [] | |
235 for i in xrange(0, test_size, batch_size): | |
236 test_batches = test_batches + \ | |
237 [(raw_test_data[i:i+batch_size], raw_test_labels[i:i+batch_size])] | |
238 | |
239 validation_batches = [] | |
240 for i in xrange(0, test_size, batch_size): | |
241 validation_batches = validation_batches + \ | |
242 [(raw_train_data[offset+i:offset+i+batch_size], raw_train_labels[offset+i:offset+i+batch_size])] | |
243 | |
244 | |
245 ishape = (32,32) # this is the size of NIST images | |
246 | |
247 # allocate symbolic variables for the data | |
248 x = T.fmatrix() # the data is presented as rasterized images | |
249 y = T.lvector() # the labels are presented as 1D vector of | |
250 # [long int] labels | |
251 | |
304
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
252 if verbose==1: |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
253 print 'finished parsing the data' |
110 | 254 # construct the logistic regression class |
255 classifier = MLP( input=x.reshape((batch_size,32*32)),\ | |
256 n_in=32*32,\ | |
257 n_hidden=nb_hidden,\ | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
258 n_out=nb_targets, |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
259 learning_rate=learning_rate) |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
260 |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
261 |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
262 |
110 | 263 |
264 # the cost we minimize during training is the negative log likelihood of | |
265 # the model plus the regularization terms (L1 and L2); cost is expressed | |
266 # here symbolically | |
267 cost = classifier.negative_log_likelihood(y) \ | |
268 + L1_reg * classifier.L1 \ | |
269 + L2_reg * classifier.L2_sqr | |
270 | |
271 # compiling a theano function that computes the mistakes that are made by | |
272 # the model on a minibatch | |
273 test_model = theano.function([x,y], classifier.errors(y)) | |
274 | |
275 # compute the gradient of cost with respect to theta = (W1, b1, W2, b2) | |
276 g_W1 = T.grad(cost, classifier.W1) | |
277 g_b1 = T.grad(cost, classifier.b1) | |
278 g_W2 = T.grad(cost, classifier.W2) | |
279 g_b2 = T.grad(cost, classifier.b2) | |
280 | |
281 # specify how to update the parameters of the model as a dictionary | |
282 updates = \ | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
283 { classifier.W1: classifier.W1 - classifier.lr*g_W1 \ |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
284 , classifier.b1: classifier.b1 - classifier.lr*g_b1 \ |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
285 , classifier.W2: classifier.W2 - classifier.lr*g_W2 \ |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
286 , classifier.b2: classifier.b2 - classifier.lr*g_b2 } |
110 | 287 |
288 # compiling a theano function `train_model` that returns the cost, but in | |
289 # the same time updates the parameter of the model based on the rules | |
290 # defined in `updates` | |
291 train_model = theano.function([x, y], cost, updates = updates ) | |
292 n_minibatches = len(train_batches) | |
293 | |
294 | |
295 | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
296 |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
297 |
110 | 298 |
299 #conditions for stopping the adaptation: | |
300 #1) we have reached nb_max_exemples (this is rounded up to be a multiple of the train size) | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
301 #2) validation error is going up twice in a row(probable overfitting) |
110 | 302 |
303 # This means we no longer stop on slow convergence as low learning rates stopped | |
304 # too fast. | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
305 |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
306 # no longer relevant |
110 | 307 patience =nb_max_exemples/batch_size |
308 patience_increase = 2 # wait this much longer when a new best is | |
309 # found | |
310 improvement_threshold = 0.995 # a relative improvement of this much is | |
311 # considered significant | |
312 validation_frequency = n_minibatches/4 | |
313 | |
314 | |
315 | |
316 | |
317 best_params = None | |
318 best_validation_loss = float('inf') | |
319 best_iter = 0 | |
320 test_score = 0. | |
321 start_time = time.clock() | |
322 n_iter = nb_max_exemples/batch_size # nb of max times we are allowed to run through all exemples | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
323 n_iter = n_iter/n_minibatches + 1 #round up |
110 | 324 n_iter=max(1,n_iter) # run at least once on short debug call |
212
e390b0454515
added classic lr time decay and py code to calculate the error based on a saved model
xaviermuller
parents:
169
diff
changeset
|
325 time_n=0 #in unit of exemples |
e390b0454515
added classic lr time decay and py code to calculate the error based on a saved model
xaviermuller
parents:
169
diff
changeset
|
326 |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
327 |
110 | 328 |
304
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
329 if verbose == 1: |
110 | 330 print 'looping at most %d times through the data set' %n_iter |
331 for iter in xrange(n_iter* n_minibatches): | |
332 | |
333 # get epoch and minibatch index | |
334 epoch = iter / n_minibatches | |
335 minibatch_index = iter % n_minibatches | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
336 |
212
e390b0454515
added classic lr time decay and py code to calculate the error based on a saved model
xaviermuller
parents:
169
diff
changeset
|
337 |
e390b0454515
added classic lr time decay and py code to calculate the error based on a saved model
xaviermuller
parents:
169
diff
changeset
|
338 if adaptive_lr==2: |
e390b0454515
added classic lr time decay and py code to calculate the error based on a saved model
xaviermuller
parents:
169
diff
changeset
|
339 classifier.lr.value = tau*initial_lr/(tau+time_n) |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
340 |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
341 |
110 | 342 # get the minibatches corresponding to `iter` modulo |
343 # `len(train_batches)` | |
344 x,y = train_batches[ minibatch_index ] | |
345 # convert to float | |
346 x_float = x/255.0 | |
347 cost_ij = train_model(x_float,y) | |
348 | |
349 if (iter+1) % validation_frequency == 0: | |
350 # compute zero-one loss on validation set | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
351 |
110 | 352 this_validation_loss = 0. |
353 for x,y in validation_batches: | |
354 # sum up the errors for each minibatch | |
355 x_float = x/255.0 | |
356 this_validation_loss += test_model(x_float,y) | |
357 # get the average by dividing with the number of minibatches | |
358 this_validation_loss /= len(validation_batches) | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
359 #save the validation loss |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
360 total_validation_error_list.append(this_validation_loss) |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
361 |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
362 #get the training error rate |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
363 this_train_loss=0 |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
364 for x,y in train_batches: |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
365 # sum up the errors for each minibatch |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
366 x_float = x/255.0 |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
367 this_train_loss += test_model(x_float,y) |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
368 # get the average by dividing with the number of minibatches |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
369 this_train_loss /= len(train_batches) |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
370 #save the validation loss |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
371 total_train_error_list.append(this_train_loss) |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
372 if(this_train_loss<best_training_error): |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
373 best_training_error=this_train_loss |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
374 |
304
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
375 if verbose == 1: |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
376 print('epoch %i, minibatch %i/%i, validation error %f, training error %f %%' % \ |
110 | 377 (epoch, minibatch_index+1, n_minibatches, \ |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
378 this_validation_loss*100.,this_train_loss*100)) |
212
e390b0454515
added classic lr time decay and py code to calculate the error based on a saved model
xaviermuller
parents:
169
diff
changeset
|
379 print 'learning rate = %f' %classifier.lr.value |
e390b0454515
added classic lr time decay and py code to calculate the error based on a saved model
xaviermuller
parents:
169
diff
changeset
|
380 print 'time = %i' %time_n |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
381 |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
382 |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
383 #save the learning rate |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
384 learning_rate_list.append(classifier.lr.value) |
110 | 385 |
386 | |
387 # if we got the best validation score until now | |
388 if this_validation_loss < best_validation_loss: | |
389 # save best validation score and iteration number | |
390 best_validation_loss = this_validation_loss | |
391 best_iter = iter | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
392 # reset patience if we are going down again |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
393 # so we continue exploring |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
394 patience=nb_max_exemples/batch_size |
110 | 395 # test it on the test set |
396 test_score = 0. | |
397 for x,y in test_batches: | |
398 x_float=x/255.0 | |
399 test_score += test_model(x_float,y) | |
400 test_score /= len(test_batches) | |
304
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
401 if verbose == 1: |
110 | 402 print((' epoch %i, minibatch %i/%i, test error of best ' |
403 'model %f %%') % | |
404 (epoch, minibatch_index+1, n_minibatches, | |
405 test_score*100.)) | |
406 | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
407 # if the validation error is going up, we are overfitting (or oscillating) |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
408 # stop converging but run at least to next validation |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
409 # to check overfitting or ocsillation |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
410 # the saved weights of the model will be a bit off in that case |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
411 elif this_validation_loss >= best_validation_loss: |
110 | 412 #calculate the test error at this point and exit |
413 # test it on the test set | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
414 # however, if adaptive_lr is true, try reducing the lr to |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
415 # get us out of an oscilliation |
145
8ceaaf812891
changed adaptive lr flag from bool to int for jobman issues
XavierMuller
parents:
143
diff
changeset
|
416 if adaptive_lr==1: |
304
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
417 classifier.lr.value=classifier.lr.value*lr_t2_factor |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
418 |
110 | 419 test_score = 0. |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
420 #cap the patience so we are allowed one more validation error |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
421 #calculation before aborting |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
422 patience = iter+validation_frequency+1 |
110 | 423 for x,y in test_batches: |
424 x_float=x/255.0 | |
425 test_score += test_model(x_float,y) | |
426 test_score /= len(test_batches) | |
304
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
427 if verbose == 1: |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
428 print ' validation error is going up, possibly stopping soon' |
110 | 429 print((' epoch %i, minibatch %i/%i, test error of best ' |
430 'model %f %%') % | |
431 (epoch, minibatch_index+1, n_minibatches, | |
432 test_score*100.)) | |
433 | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
434 |
110 | 435 |
436 | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
437 if iter>patience: |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
438 print 'we have diverged' |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
439 break |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
440 |
110 | 441 |
212
e390b0454515
added classic lr time decay and py code to calculate the error based on a saved model
xaviermuller
parents:
169
diff
changeset
|
442 time_n= time_n + batch_size |
110 | 443 end_time = time.clock() |
304
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
444 if verbose == 1: |
110 | 445 print(('Optimization complete. Best validation score of %f %% ' |
446 'obtained at iteration %i, with test performance %f %%') % | |
447 (best_validation_loss * 100., best_iter, test_score*100.)) | |
448 print ('The code ran for %f minutes' % ((end_time-start_time)/60.)) | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
449 print iter |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
450 |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
451 #save the model and the weights |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
452 numpy.savez('model.npy', config=configuration, W1=classifier.W1.value,W2=classifier.W2.value, b1=classifier.b1.value,b2=classifier.b2.value) |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
453 numpy.savez('results.npy',config=configuration,total_train_error_list=total_train_error_list,total_validation_error_list=total_validation_error_list,\ |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
454 learning_rate_list=learning_rate_list) |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
455 |
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
456 return (best_training_error*100.0,best_validation_loss * 100.,test_score*100.,best_iter*batch_size,(end_time-start_time)/60) |
110 | 457 |
458 | |
459 if __name__ == '__main__': | |
460 mlp_full_mnist() | |
461 | |
462 def jobman_mlp_full_nist(state,channel): | |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
463 (train_error,validation_error,test_error,nb_exemples,time)=mlp_full_nist(learning_rate=state.learning_rate,\ |
304
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
464 nb_max_exemples=state.nb_max_exemples,\ |
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
465 nb_hidden=state.nb_hidden,\ |
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
466 adaptive_lr=state.adaptive_lr,\ |
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
467 tau=state.tau,\ |
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
468 verbose = state.verbose,\ |
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
469 train_data = state.train_data,\ |
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
470 train_labels = state.train_labels,\ |
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
471 test_data = state.test_data,\ |
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
472 test_labels = state.test_labels,\ |
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
473 lr_t2_factor=state.lr_t2_factor) |
143
f341a4efb44a
added adaptive lr, weight file save, traine error and error curves
XavierMuller
parents:
110
diff
changeset
|
474 state.train_error=train_error |
110 | 475 state.validation_error=validation_error |
476 state.test_error=test_error | |
477 state.nb_exemples=nb_exemples | |
478 state.time=time | |
304
1e4bf5a5b46d
added type 2 adaptive learning configurable learning weight + versionning
xaviermuller
parents:
237
diff
changeset
|
479 pylearn.version.record_versions(state,[theano,ift6266,pylearn]) |
110 | 480 return channel.COMPLETE |
481 | |
482 |