diff options
Diffstat (limited to 'layer/init.lua')
-rw-r--r-- | layer/init.lua | 23 |
1 files changed, 23 insertions, 0 deletions
diff --git a/layer/init.lua b/layer/init.lua index a98621d..c8c691b 100644 --- a/layer/init.lua +++ b/layer/init.lua @@ -44,3 +44,26 @@ end function nerv.Layer:back_propagate(next_bp_err, bp_err, input, output) nerv.error_method_not_implemented() end + +function nerv.Layer:check_dim_len(len_in, len_out) + local expected_in = #self.dim_in + local expected_out = #self.dim_out + if len_in > 0 and expected_in ~= len_in then + nerv.error("layer %s expects %d inputs, %d given", + self.id, len_in, expected_in) + end + if len_out > 0 and expected_out ~= len_out then + nerv.error("layer %s expects %d outputs, %d given", + self.id, len_out, expected_out) + end +end + +function nerv.Layer:get_dim() + return self.dim_in, self.dim_out +end + +require 'layer.affine' +require 'layer.sigmoid' +require 'layer.softmax_ce' +require 'layer.bias' +require 'layer.window' |