diff options
author | txh18 <[email protected]> | 2015-11-11 10:41:14 +0800 |
---|---|---|
committer | txh18 <[email protected]> | 2015-11-11 10:41:14 +0800 |
commit | 5761e68ec1b73ed867443fb2687739395f22f2f9 (patch) | |
tree | ef7bde16e3faeb6eb1bbedc5258b6a1ae9d89952 /nerv/layer/affine.lua | |
parent | 8f8ad5da3d178bdd74c76cea37946f5c950ed5c0 (diff) |
got good result when batch_size=1, strange!
Diffstat (limited to 'nerv/layer/affine.lua')
-rw-r--r-- | nerv/layer/affine.lua | 9 |
1 files changed, 8 insertions, 1 deletions
diff --git a/nerv/layer/affine.lua b/nerv/layer/affine.lua index 015ec3f..0462383 100644 --- a/nerv/layer/affine.lua +++ b/nerv/layer/affine.lua @@ -31,7 +31,14 @@ function LinearTransParam:update(gradient) MatrixParam.update(self, gradient) local gconf = self.gconf -- weight decay - self.trans:add(self.trans, self.trans, 1.0, -gconf.lrate * gconf.wcost) + self.trans:add(self.trans, self.trans, 1.0, -gconf.lrate * gconf.wcost / gconf.batch_size) +end + +function BiasParam:update(gradient) + MatrixParam.update(self, gradient) + local gconf = self.gconf + -- weight decay + self.trans:add(self.trans, self.trans, 1.0, -gconf.lrate * gconf.wcost / gconf.batch_size) end function AffineLayer:__init(id, global_conf, layer_conf) |