Mercurial > ift6266
changeset 286:1cc535f3e254
correction d'un bug pour affichage des resultats de pre-train avec P07
author | SylvainPL <sylvain.pannetier.lebeuf@umontreal.ca> |
---|---|
date | Thu, 25 Mar 2010 12:20:27 -0400 |
parents | 694e75413413 |
children | f9b93ae45723 80ee63c3e749 |
files | deep/stacked_dae/v_sylvain/sgd_optimization.py |
diffstat | 1 files changed, 3 insertions(+), 1 deletions(-) [+] |
line wrap: on
line diff
--- a/deep/stacked_dae/v_sylvain/sgd_optimization.py Wed Mar 24 15:14:24 2010 -0400 +++ b/deep/stacked_dae/v_sylvain/sgd_optimization.py Thu Mar 25 12:20:27 2010 -0400 @@ -87,6 +87,8 @@ def pretrain(self,dataset): print "STARTING PRETRAINING, time = ", datetime.datetime.now() sys.stdout.flush() + + un_fichier=int(819200.0/self.hp.minibatch_size) #Number of batches in a P07 batch start_time = time.clock() ## Pre-train layer-wise @@ -113,7 +115,7 @@ #When we pass through the data only once (the case with P07) #There is approximately 800*1024=819200 examples per file (1k per example and files are 800M) - if self.hp.pretraining_epochs_per_layer == 1 and count%819200 == 0: + if self.hp.pretraining_epochs_per_layer == 1 and count%un_fichier == 0: print 'Pre-training layer %i, epoch %d, cost '%(i,num_files),c num_files+=1 sys.stdout.flush()