diff options
author | txh18 <cloudygooseg@gmail.com> | 2016-02-16 19:42:29 +0800 |
---|---|---|
committer | txh18 <cloudygooseg@gmail.com> | 2016-02-16 19:42:29 +0800 |
commit | 490a10c2130773bd022f05513fa2905b6a6c6e91 (patch) | |
tree | a66d613c3c6cfc4054e799606f7fb56278281347 /nerv/layer/init.lua | |
parent | 1721de3a5f5cd6df731484a8161b537468bea0bd (diff) |
fixed some minor problem
Diffstat (limited to 'nerv/layer/init.lua')
-rw-r--r-- | nerv/layer/init.lua | 1 |
1 files changed, 1 insertions, 0 deletions
diff --git a/nerv/layer/init.lua b/nerv/layer/init.lua index d952022..43c2250 100644 --- a/nerv/layer/init.lua +++ b/nerv/layer/init.lua @@ -91,6 +91,7 @@ function Layer:find_param(pid_list, lconf, gconf, p_type, p_dim) end nerv.info("param [%s] of layer [%s] is not found in `layer_conf` or `layer_conf.pr`, " .. "switch to auto-generate.", pid_list_str, self.id) + local pid_g = self.id .. '_' .. pid_list[1] p = p_type(pid_g, gconf) p.trans = gconf.cumat_type(unpack(p_dim)) if type(gconf.param_random) ~= "function" then |