# HG changeset patch # User Frederic Bastien # Date 1264604197 18000 # Node ID 912be602c3ac0909435bcae8d22b841ca876052f # Parent a13142cbeabdc3d803183526def13f84899ca4f1 fix import for the move of shared var outside of the sandbox. diff -r a13142cbeabd -r 912be602c3ac pylearn/dataset_ops/gldataset.py --- a/pylearn/dataset_ops/gldataset.py Thu Jan 21 13:17:09 2010 -0500 +++ b/pylearn/dataset_ops/gldataset.py Wed Jan 27 09:56:37 2010 -0500 @@ -17,8 +17,8 @@ import numpy import theano -from theano.compile.sandbox import shared -from theano.compile.sandbox import pfunc as function +from theano.compile import shared +from theano.compile import pfunc as function _logger = logging.getLogger('gldataset') def debug(*msg): _logger.debug(' '.join(str(m) for m in msg)) diff -r a13142cbeabd -r 912be602c3ac pylearn/dataset_ops/tests/test_cifar10.py --- a/pylearn/dataset_ops/tests/test_cifar10.py Thu Jan 21 13:17:09 2010 -0500 +++ b/pylearn/dataset_ops/tests/test_cifar10.py Wed Jan 27 09:56:37 2010 -0500 @@ -1,7 +1,7 @@ import unittest import numpy import theano -from theano.compile.sandbox import pfunc, shared +from theano.compile import pfunc, shared from theano import tensor from pylearn.dataset_ops.cifar10 import cifar10, forget diff -r a13142cbeabd -r 912be602c3ac pylearn/shared/layers/exponential_mean.py --- a/pylearn/shared/layers/exponential_mean.py Thu Jan 21 13:17:09 2010 -0500 +++ b/pylearn/shared/layers/exponential_mean.py Wed Jan 27 09:56:37 2010 -0500 @@ -6,7 +6,7 @@ import numpy import theano import theano.tensor -from theano.compile.sandbox import shared +from theano.compile import shared class ExponentialMean(object): """Maintain an exponentially-decaying estimate of the mean diff -r a13142cbeabd -r 912be602c3ac pylearn/shared/layers/kouh2008.py --- a/pylearn/shared/layers/kouh2008.py Thu Jan 21 13:17:09 2010 -0500 +++ b/pylearn/shared/layers/kouh2008.py Wed Jan 27 09:56:37 2010 -0500 @@ -20,7 +20,7 @@ from theano import tensor from theano.tensor.nnet import softplus from theano.sandbox.softsign import softsign -from theano.compile.sandbox import shared +from theano.compile import shared from pylearn.shared.layers.util import add_logging, update_locals try: diff -r a13142cbeabd -r 912be602c3ac pylearn/shared/layers/lecun1998.py --- a/pylearn/shared/layers/lecun1998.py Thu Jan 21 13:17:09 2010 -0500 +++ b/pylearn/shared/layers/lecun1998.py Wed Jan 27 09:56:37 2010 -0500 @@ -6,7 +6,7 @@ import theano from theano import tensor -from theano.compile.sandbox import shared, pfunc +from theano.compile import shared, pfunc from theano.sandbox.conv import ConvOp from theano.sandbox.downsample import DownsampleFactorMax diff -r a13142cbeabd -r 912be602c3ac pylearn/shared/layers/logreg.py --- a/pylearn/shared/layers/logreg.py Thu Jan 21 13:17:09 2010 -0500 +++ b/pylearn/shared/layers/logreg.py Wed Jan 27 09:56:37 2010 -0500 @@ -2,7 +2,7 @@ """ import numpy import theano -from theano.compile.sandbox import shared +from theano.compile import shared from theano.tensor import nnet from pylearn.shared.layers.util import update_locals, add_logging diff -r a13142cbeabd -r 912be602c3ac pylearn/shared/layers/rust2005.py --- a/pylearn/shared/layers/rust2005.py Thu Jan 21 13:17:09 2010 -0500 +++ b/pylearn/shared/layers/rust2005.py Wed Jan 27 09:56:37 2010 -0500 @@ -25,7 +25,7 @@ import theano import theano.tensor import theano.tensor.nnet -from theano.compile.sandbox import shared +from theano.compile import shared from theano.sandbox.softsign import softsign from theano.tensor.nnet import softplus from theano.sandbox.conv import ConvOp diff -r a13142cbeabd -r 912be602c3ac pylearn/shared/layers/sandbox/linsvm.py --- a/pylearn/shared/layers/sandbox/linsvm.py Thu Jan 21 13:17:09 2010 -0500 +++ b/pylearn/shared/layers/sandbox/linsvm.py Wed Jan 27 09:56:37 2010 -0500 @@ -1,6 +1,6 @@ import numpy import theano -from theano.compile.sandbox import shared +from theano.compile import shared from theano.tensor import nnet from .util import update_locals diff -r a13142cbeabd -r 912be602c3ac pylearn/shared/layers/sgd.py --- a/pylearn/shared/layers/sgd.py Thu Jan 21 13:17:09 2010 -0500 +++ b/pylearn/shared/layers/sgd.py Wed Jan 27 09:56:37 2010 -0500 @@ -4,7 +4,7 @@ import numpy import theano from theano import tensor -from theano.compile.sandbox import shared +from theano.compile import shared class StochasticGradientDescent(object): """Fixed stepsize gradient descent diff -r a13142cbeabd -r 912be602c3ac pylearn/shared/layers/sigmoidal_layer.py --- a/pylearn/shared/layers/sigmoidal_layer.py Thu Jan 21 13:17:09 2010 -0500 +++ b/pylearn/shared/layers/sigmoidal_layer.py Wed Jan 27 09:56:37 2010 -0500 @@ -6,7 +6,7 @@ import theano from theano import tensor -from theano.compile.sandbox import shared, pfunc +from theano.compile import shared, pfunc from pylearn.shared.layers.util import update_locals, add_logging from pylearn.shared.layers.squash import squash diff -r a13142cbeabd -r 912be602c3ac pylearn/shared/layers/tests/test_kouh2008.py --- a/pylearn/shared/layers/tests/test_kouh2008.py Thu Jan 21 13:17:09 2010 -0500 +++ b/pylearn/shared/layers/tests/test_kouh2008.py Wed Jan 27 09:56:37 2010 -0500 @@ -1,7 +1,7 @@ import numpy import theano.compile.debugmode from theano import tensor -from theano.compile.sandbox import pfunc +from theano.compile import pfunc from pylearn.shared.layers import LogisticRegression, Kouh2008 def test_dtype(): diff -r a13142cbeabd -r 912be602c3ac pylearn/shared/layers/tests/test_sigmoidal_layer.py --- a/pylearn/shared/layers/tests/test_sigmoidal_layer.py Thu Jan 21 13:17:09 2010 -0500 +++ b/pylearn/shared/layers/tests/test_sigmoidal_layer.py Wed Jan 27 09:56:37 2010 -0500 @@ -1,7 +1,7 @@ import numpy from pylearn.shared.layers import SigmoidalLayer, LogisticRegression from theano import tensor -from theano.compile.sandbox import shared, pfunc +from theano.compile import shared, pfunc def test_w_random(dtype='float64'): if dtype == 'float64':