Mercurial > pylearn
diff filetensor.py @ 33:bb92087cb0f6
added filetensor.py
author | bergstrj@iro.umontreal.ca |
---|---|
date | Thu, 17 Apr 2008 12:49:33 -0400 |
parents | |
children | 2508c373cf29 |
line wrap: on
line diff
--- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/filetensor.py Thu Apr 17 12:49:33 2008 -0400 @@ -0,0 +1,168 @@ +""" +Read and write the matrix file format described at +http://www.cs.nyu.edu/~ylclab/data/norb-v1.0/index.html + +The format is for dense tensors: + + magic number indicating type and endianness - 4bytes + rank of tensor - int32 + dimensions - int32, int32, int32, ... + <data> + +The number of dimensions and rank is slightly tricky: + for scalar: rank=0, dimensions = [1, 1, 1] + for vector: rank=1, dimensions = [?, 1, 1] + for matrix: rank=2, dimensions = [?, ?, 1] + +For rank >= 3, the number of dimensions matches the rank exactly. + +""" +import sys +import numpy + +def prod(lst): + p = 1 + for l in lst: + p *= l + return p + +_magic_dtype = { + 0x1E3D4C51 : ('float32', 4), + 0x1E3D4C52 : ('packed matrix', 0), #what is a packed matrix? + 0x1E3D4C53 : ('float64', 8), + 0x1E3D4C54 : ('int32', 4), + 0x1E3D4C55 : ('int8', 1), + 0x1E3D4C56 : ('int16', 2), + } +_dtype_magic = { + 'float32': 0x1E3D4C51, + 'packed matrix': 0x1E3D4C52, + 'float64': 0x1E3D4C53, + 'int32': 0x1E3D4C54, + 'int8': 0x1E3D4C55, + 'int16': 0x1E3D4C56 + } + +def _unused(): + f.seek(0,2) #seek to end + f_len = f.tell() + f.seek(f_data_start,0) #seek back to where we were + + if debug: print 'length:', f_len + + + f_data_bytes = (f_len - f_data_start) + + if debug: print 'data bytes according to header: ', dim_size * elsize + if debug: print 'data bytes according to file : ', f_data_bytes + + if debug: print 'reading data...' + sys.stdout.flush() + +def _write_int32(f, i): + i_array = numpy.asarray(i, dtype='int32') + if 0: print 'writing int32', i, i_array + i_array.tofile(f) +def _read_int32(f): + s = f.read(4) + s_array = numpy.fromstring(s, dtype='int32') + return s_array.item() + +def read_ndarray(f, dim, dtype): + return numpy.fromfile(f, dtype=dtype, count=prod(dim)).reshape(dim) + +# +# TODO: implement item selection: +# e.g. load('some mat', subtensor=(:6, 2:5)) +# +# This function should be memory efficient by: +# - allocating an output matrix at the beginning +# - seeking through the file, reading subtensors from multiple places +def read(f, subtensor=None, debug=False): + """Load all or part of file 'f' into a numpy ndarray + + If f is a string, it will be treated as a filename, and opened in read mode. + + If subtensor is not None, it should be like the argument to + numpy.ndarray.__getitem__. The following two expressions should return + equivalent ndarray objects, but the one on the left may be faster and more + memory efficient if the underlying file f is big. + + read(f, subtensor) <===> read(f)[*subtensor] + + Support for subtensors is currently spotty, so check the code to see if your + particular type of subtensor is supported. + + """ + + if isinstance(f, str): + if debug: print 'f', f + f = file(f, 'r') + + #what is the data type of this matrix? + #magic_s = f.read(4) + #magic = numpy.fromstring(magic_s, dtype='int32') + magic = _read_int32(f) + magic_t, elsize = _magic_dtype[magic] + if debug: + print 'header magic', magic, magic_t, elsize + if magic_t == 'packed matrix': + raise NotImplementedError('packed matrix not supported') + + #what is the rank of the tensor? + ndim = _read_int32(f) + if debug: print 'header ndim', ndim + + #what are the dimensions of the tensor? + dim = numpy.fromfile(f, dtype='int32', count=max(ndim,3))[:ndim] + dim_size = prod(dim) + if debug: print 'header dim', dim, dim_size + + rval = None + if subtensor is None: + rval = read_ndarray(f, dim, magic_t) + elif isinstance(subtensor, slice): + if subtensor.step not in (None, 1): + raise NotImplementedError('slice with step', subtensor.step) + if subtensor.start not in (None, 0): + bytes_per_row = prod(dim[1:]) * elsize + raise NotImplementedError('slice with start', subtensor.start) + dim[0] = min(dim[0], subtensor.stop) + rval = read_ndarray(f, dim, magic_t) + else: + raise NotImplementedError('subtensor access not written yet:', subtensor) + + return rval + +def write(f, mat): + if isinstance(f, str): + f = file(f, 'w') + + _write_int32(f, _dtype_magic[str(mat.dtype)]) + _write_int32(f, len(mat.shape)) + shape = mat.shape + if len(shape) < 3: + shape = list(shape) + [1] * (3 - len(shape)) + print 'writing shape =', shape + for sh in shape: + _write_int32(f, sh) + mat.tofile(f) + +if __name__ == '__main__': + #a small test script, starts by reading sys.argv[1] + rval = read(sys.argv[1], slice(400), debug=True) #load from filename + print 'rval', rval.shape, rval.size + + f = file('/tmp/some_mat', 'w'); + write(f, rval) + print '' + f.close() + f = file('/tmp/some_mat', 'r'); + rval2 = read(f) #load from file handle + print 'rval2', rval2.shape, rval2.size + + assert rval.dtype == rval2.dtype + assert rval.shape == rval2.shape + assert numpy.all(rval == rval2) + print 'ok' +