@@ -74,10 +74,10 @@ def create_net(mini_batch_size=10, activation_fn=tanh):
74
74
#ConvPoolLayer(image_shape=(mini_batch_size, 20, 12, 12), filter_shape=(40, 20, 5, 5), poolsize=(2, 2), activation_fn=activation_fn),
75
75
#FullyConnectedLayer(n_in=40*4*4, n_out=100, mini_batch_size=mini_batch_size, activation_fn=activation_fn),
76
76
#FullyConnectedLayer(n_in=784, n_out=100, mini_batch_size=mini_batch_size, activation_fn=activation_fn),
77
- # FullyConnectedLayer(n_in=20*12*12, n_out=100, mini_batch_size=mini_batch_size ),
77
+ FullyConnectedLayer (n_in = 20 * 12 * 12 , n_out = 100 ),
78
78
#FullyConnectedLayer(n_in=100, n_out=100, mini_batch_size=mini_batch_size, activation_fn=activation_fn),
79
- # SoftmaxLayer(n_in=100, n_out=10, mini_batch_size=mini_batch_size )], mini_batch_size)
80
- SoftmaxLayer (n_in = 20 * 12 * 12 , n_out = 10 )], mini_batch_size )
79
+ SoftmaxLayer (n_in = 100 , n_out = 10 )], mini_batch_size )
80
+ # SoftmaxLayer(n_in=20*12*12, n_out=10)], mini_batch_size)
81
81
82
82
#### Load the MNIST data
83
83
def load_data_shared (filename = "../data/mnist.pkl.gz" ):
@@ -252,7 +252,7 @@ def set_inpt(self, inpt, mini_batch_size):
252
252
253
253
class FullyConnectedLayer ():
254
254
255
- def __init__ (self , n_in , n_out , mini_batch_size = 10 , activation_fn = sigmoid ):
255
+ def __init__ (self , n_in , n_out , activation_fn = sigmoid ):
256
256
self .n_in = n_in
257
257
self .n_out = n_out
258
258
self .activation_fn = activation_fn
0 commit comments