aboutsummaryrefslogtreecommitdiff
path: root/layer/sigmoid.lua
diff options
context:
space:
mode:
authorDeterminant <ted.sybil@gmail.com>2015-06-02 12:51:18 +0800
committerDeterminant <ted.sybil@gmail.com>2015-06-02 12:51:18 +0800
commit0d3d8f4afdc38726b8ed933dbfcb85e759145c43 (patch)
treed0ea9b021e710b9ac8aea4bbcd56922f3fe1f1fe /layer/sigmoid.lua
parentbf05d75bf173e1a496a277c76593537dc9cdb28a (diff)
add preprocessing layers and change layer constructor interface
Diffstat (limited to 'layer/sigmoid.lua')
-rw-r--r--layer/sigmoid.lua12
1 files changed, 9 insertions, 3 deletions
diff --git a/layer/sigmoid.lua b/layer/sigmoid.lua
index ca34419..220b7af 100644
--- a/layer/sigmoid.lua
+++ b/layer/sigmoid.lua
@@ -1,11 +1,17 @@
local SigmoidLayer = nerv.class("nerv.SigmoidLayer", "nerv.Layer")
-function SigmoidLayer:__init(id, global_conf)
+function SigmoidLayer:__init(id, global_conf, layer_conf)
self.id = id
self.gconf = global_conf
+ self.dim_in = layer_conf.dim_in
+ self.dim_out = layer_conf.dim_out
+ self:check_dim_len(1, 1)
end
function SigmoidLayer:init()
+ if self.dim_in[1] ~= self.dim_out[1] then
+ nerv.error("mismatching dimensions of input and output")
+ end
end
function SigmoidLayer:update(bp_err, input, output)
@@ -13,9 +19,9 @@ function SigmoidLayer:update(bp_err, input, output)
end
function SigmoidLayer:propagate(input, output)
- output[0]:sigmoid(input[0])
+ output[1]:sigmoid(input[1])
end
function SigmoidLayer:back_propagate(next_bp_err, bp_err, input, output)
- next_bp_err[0]:sigmoid_grad(bp_err[0], output[0])
+ next_bp_err[1]:sigmoid_grad(bp_err[1], output[1])
end