@@ -61,10 +61,6 @@ def unzip(zipped):
6161 return new_params
6262
6363
64- def itemlist (tparams ):
65- return [vv for kk , vv in tparams .iteritems ()]
66-
67-
6864def dropout_layer (state_before , use_noise , trng ):
6965 proj = tensor .switch (use_noise ,
7066 (state_before *
@@ -232,7 +228,7 @@ def adadelta(lr, tparams, grads, x, mask, y, cost):
232228 running_grads2 )]
233229 ru2up = [(ru2 , 0.95 * ru2 + 0.05 * (ud ** 2 ))
234230 for ru2 , ud in zip (running_up2 , updir )]
235- param_up = [(p , p + ud ) for p , ud in zip (itemlist ( tparams ), updir )]
231+ param_up = [(p , p + ud ) for p , ud in zip (tparams . values ( ), updir )]
236232
237233 f_update = theano .function ([lr ], [], updates = ru2up + param_up ,
238234 on_unused_input = 'ignore' )
@@ -266,7 +262,7 @@ def rmsprop(lr, tparams, grads, x, mask, y, cost):
266262 for ud , zg , rg , rg2 in zip (updir , zipped_grads , running_grads ,
267263 running_grads2 )]
268264 param_up = [(p , p + udn [1 ])
269- for p , udn in zip (itemlist ( tparams ), updir_new )]
265+ for p , udn in zip (tparams . values ( ), updir_new )]
270266 f_update = theano .function ([lr ], [], updates = updir_new + param_up ,
271267 on_unused_input = 'ignore' )
272268
@@ -280,7 +276,7 @@ def sgd(lr, tparams, grads, x, mask, y, cost):
280276
281277 f_grad_shared = theano .function ([x , mask , y ], cost , updates = gsup )
282278
283- pup = [(p , p - lr * g ) for p , g in zip (itemlist ( tparams ), gshared )]
279+ pup = [(p , p - lr * g ) for p , g in zip (tparams . values ( ), gshared )]
284280 f_update = theano .function ([lr ], [], updates = pup )
285281
286282 return f_grad_shared , f_update
0 commit comments