Mercurial > ift6266
view datasets/dataset.py @ 186:d364a130b221
Ajout du code de base pour scalar_series. Modifications à stacked_dae: réglé un problème avec les input_divider (empêchait une optimisation), et ajouté utilisation des séries. Si j'avais pas déjà commité, aussi, j'ai enlevé l'histoire de réutilisation du pretraining: c'était compliqué (error prone) et ça créait des jobs beaucoup trop longues.
author | fsavard |
---|---|
date | Mon, 01 Mar 2010 11:45:25 -0500 |
parents | 4b28d7382dbf |
children | d6672a7daea5 |
line wrap: on
line source
from dsetiter import DataIterator class DataSet(object): def test(self, batchsize, bufsize=None): r""" Returns an iterator over the test examples. Parameters batchsize (int) -- the size of the minibatches, 0 means return the whole set at once. bufsize (int, optional) -- the size of the in-memory buffer, 0 to disable. """ return self._return_it(batchsize, bufsize, self._test) def train(self, batchsize, bufsize=None): r""" Returns an iterator over the training examples. Parameters batchsize (int) -- the size of the minibatches, 0 means return the whole set at once. bufsize (int, optional) -- the size of the in-memory buffer, 0 to disable. """ return self._return_it(batchsize, bufsize, self._train) def valid(self, batchsize, bufsize=None): r""" Returns an iterator over the validation examples. Parameters batchsize (int) -- the size of the minibatches, 0 means return the whole set at once. bufsize (int, optional) -- the size of the in-memory buffer, 0 to disable. """ return self._return_it(batchsize, bufsize, self._valid) def _return_it(batchsize, bufsize, data): r""" Must return an iterator over the specified dataset (`data`). Implement this in subclassses. """ raise NotImplemented