aboutsummaryrefslogtreecommitdiff
path: root/layer/softmax_ce.lua
diff options
context:
space:
mode:
authorcloudygoose <[email protected]>2015-06-03 10:29:41 +0800
committercloudygoose <[email protected]>2015-06-03 10:29:41 +0800
commitbf01fd6cea42def51becb6ea866d4fd335e45842 (patch)
tree09d12e50e3a6156c7e0cd7412b22fa4b61189495 /layer/softmax_ce.lua
parent6984519cbb659aac0b0b323de93d5a90aa2049b7 (diff)
parentbb56a806e0636a0b20117b1644701d63e2bfaefb (diff)
...
Merge remote-tracking branch 'upstream/master'
Diffstat (limited to 'layer/softmax_ce.lua')
-rw-r--r--layer/softmax_ce.lua18
1 files changed, 13 insertions, 5 deletions
diff --git a/layer/softmax_ce.lua b/layer/softmax_ce.lua
index 37d2864..09eb3a9 100644
--- a/layer/softmax_ce.lua
+++ b/layer/softmax_ce.lua
@@ -1,11 +1,17 @@
local SoftmaxCELayer = nerv.class("nerv.SoftmaxCELayer", "nerv.Layer")
-function SoftmaxCELayer:__init(id, global_conf)
+function SoftmaxCELayer:__init(id, global_conf, layer_conf)
self.id = id
self.gconf = global_conf
+ self.dim_in = layer_conf.dim_in
+ self.dim_out = layer_conf.dim_out
+ self:check_dim_len(2, -1) -- two inputs: nn output and label
end
function SoftmaxCELayer:init()
+ if self.dim_in[1] ~= self.dim_in[1] then
+ nerv.error("mismatching dimensions of previous network output and labels")
+ end
self.total_ce = 0.0
self.total_frames = 0
end
@@ -15,12 +21,14 @@ function SoftmaxCELayer:update(bp_err, input, output)
end
function SoftmaxCELayer:propagate(input, output)
- local soutput = input[0]:create() -- temporary value for calc softmax
+ local soutput = input[1]:create() -- temporary value for calc softmax
self.soutput = soutput
- soutput:softmax(input[0])
+ soutput:softmax(input[1])
local ce = soutput:create()
ce:log_elem(soutput)
- ce:mul_elem(ce, input[1])
+ ce:mul_elem(ce, input[2])
+-- print(input[1][0])
+-- print(soutput[1][0])
-- add total ce
self.total_ce = self.total_ce - ce:rowsum():colsum()[0]
self.total_frames = self.total_frames + soutput:nrow()
@@ -28,5 +36,5 @@ end
function SoftmaxCELayer:back_propagate(next_bp_err, bp_err, input, output)
-- softmax output - label
- next_bp_err[0]:add(self.soutput, input[1], 1.0, -1.0)
+ next_bp_err[1]:add(self.soutput, input[2], 1.0, -1.0)
end