Mercurial > pylearn
view examples/linear_classifier.py @ 485:e8c37244b54f
Small bugfix in regularization
author | Joseph Turian <turian@gmail.com> |
---|---|
date | Tue, 28 Oct 2008 01:37:32 -0400 |
parents | 4060812caa22 |
children |
line wrap: on
line source
#! /usr/bin/env python """ T. Bertin-Mahieux (2008) University of Montreal bertinmt@iro.umontreal.ca linear_classifier.py Simple script that creates a linear_classifier, and learns the parameters using backpropagation. This is to illustrate how to use theano/pylearn. Anyone who knows how to make this script simpler/clearer is welcome to make the modifications. """ import os import sys import time import copy import pickle import numpy import numpy as N import numpy.random as NR from pylearn import cost import theano from theano import tensor as T def cost_function(*args,**kwargs) : """ default cost function, quadratic """ return cost.quadratic(*args,**kwargs) class modelgraph() : """ class that contains the graph of the model """ lr = T.scalar() # learning rate inputs = T.matrix() # inputs (one example per line) true_outputs = T.matrix() # outputs (one example per line) W = T.matrix() # weights input * W + b= output b = T.vector() # bias outputs = T.dot(inputs,W) + b # output, one per line costs = cost_function(true_outputs,outputs) # costs g_W = T.grad(costs,W) # gradient of W g_b = T.grad(costs,b) # gradient of b new_W = T.sub_inplace(W, lr * g_W) # update inplace of W new_b = T.sub_inplace(b, lr * g_b) # update inplace of b class model() : """ The model! Contains needed matrices, needed functions, and a link to the model graph. """ def __init__(self,input_size,output_size) : """ init matrix and bias, creates the graph, create a dict of compiled functions """ # graph self.graph = modelgraph() # weights and bias, saved in self.params seed = 666 r = NR.RandomState(seed) W = r.uniform(size = [input_size, output_size], low = -1/N.sqrt(input_size), high = 1/N.sqrt(input_size)) b = numpy.zeros((output_size, )) self.params = [W,b] # dictionary of compiled functions self.func_dict = dict() # keep some init_infos (may not be necessary) self.init_params = [input_size,output_size] def update(self,lr,true_inputs,true_outputs) : """ does an update of the model, one gradient descent """ # do we already have the proper theano function? if self.func_dict.has_key('update_func') : self.func_dict['update_func'](lr,true_inputs,true_outputs,self.params[0],self.params[1]) return else : # create the theano function, tell him what are the inputs and outputs) func = theano.function([self.graph.lr,self.graph.inputs,self.graph.true_outputs, self.graph.W, self.graph.b], [self.graph.new_W,self.graph.new_b]) # add function to dictionary, so we don't compile it again self.func_dict['update_func'] = func # use this function func(lr,true_inputs,true_outputs,self.params[0],self.params[1]) return def costs(self,true_inputs,true_outputs) : """ get the costs for given examples, don't update """ # do we already have the proper theano function? if self.func_dict.has_key('costs_func') : return self.func_dict['costs_func'](true_inputs,true_outputs,self.params[0],self.params[1]) else : # create the theano function, tell him what are the inputs and outputs) func = theano.function([self.graph.inputs,self.graph.true_outputs,self.graph.W,self.graph.b], [self.graph.costs]) # add function to dictionary, se we don't compile it again self.func_dict['costs_func'] = func # use this function return func(true_inputs,true_outputs,self.params[0],self.params[1]) def outputs(self,true_inputs) : """ get the output for a set of examples (could be called 'predict') """ # do we already have the proper theano function? if self.func_dict.has_key('outputs_func') : return self.func_dict['outputs_func'](true_inputs,self.params[0],self.params[1]) else : # create the theano function, tell him what are the inputs and outputs) func = theano.function([self.graph.inputs, self.graph.W, self.graph.b], [self.graph.outputs]) # add function to dictionary, se we don't compile it again self.func_dict['outputs_func'] = func # use this function return func(true_inputs,self.params[0],self.params[1]) def __getitem__(self,inputs) : """ for simplicity, we can use the model this way: predictions = model[inputs] """ return self.outputs(inputs) def __getstate__(self) : """ To save/copy the model, used by pickle.dump() and by copy.deepcopy(). @return a dictionnary with the params (matrix + bias) """ d = dict() d['params'] = self.params d['init_params'] = self.init_params return d def __setstate__(self,d) : """ Get the dictionary created by __getstate__(), use it to recreate the model. """ self.params = d['params'] self.init_params = d['init_params'] self.graph = modelgraph() # we did not save the model graph def __str__(self) : """ returns a string representing the model """ res = "Linear regressor, input size =",str(self.init_params[0]) res += ", output size =", str(self.init_params[1]) return res def __equal__(self,other) : """ Compares the model based on the params. @return True if the params are the same, False otherwise """ # class if not isinstance(other,model) : return False # input size if self.params[0].shape[0] != other.params[0].shape[0] : return False # output size if self.params[0].shape[1] != other.params[0].shape[1] : return False # actual values if not (self.params[0] == other.params[0]).all(): return False if not (self.params[1] == other.params[1]).all(): return False # all good return True def die_with_usage() : """ help menu """ print 'simple script to illustrate how to use theano/pylearn' print 'to launch:' print ' python linear_classifier.py -launch' sys.exit(0) #************************************************************ # main if __name__ == '__main__' : if len(sys.argv) < 2 : die_with_usage() # print create data inputs = numpy.array([[.1,.2], [.2,.8], [.9,.3], [.6,.5]]) outputs = numpy.array([[0], [0], [1], [1]]) assert inputs.shape[0] == outputs.shape[0] # create model m = model(2,1) # predict print 'prediction before training:' print m[inputs] # update it for 100 iterations for k in range(50) : m.update(.1,inputs,outputs) # predict print 'prediction after training:' print m[inputs] # show points import pylab as P colors = outputs.flatten().tolist() x = inputs[:,0] y = inputs[:,1] P.plot(x[numpy.where(outputs==0)[0]],y[numpy.where(outputs==0)[0]],'r+') P.plot(x[numpy.where(outputs==1)[0]],y[numpy.where(outputs==1)[0]],'b+') # decision line p1 = (.5 - m.params[1] * 1.) / m.params[0][1,0] # abs = 0 p2 = (.5 - m.params[1] * 1.) / m.params[0][0,0] # ord = 0 P.plot((0,p2[0],2*p2[0]),(p1[0],0,-p1[0]),'g-') # show P.axis([-1,2,-1,2]) P.show()