aboutsummaryrefslogtreecommitdiff
path: root/nerv/matrix
diff options
context:
space:
mode:
authormfy43 <ted.sybil@gmail.com>2016-04-30 14:48:29 +0800
committermfy43 <ted.sybil@gmail.com>2016-04-30 14:48:29 +0800
commit1aaffa6b9a56d4d580dea7bf3f1b7df1eec5da2f (patch)
treed705bbec5b701774a144810c5e60a73e346f9c39 /nerv/matrix
parent6051c37a2c55365a5834d5b4c11d973e2e4e1ad4 (diff)
parent1f5568a9e1457dcf5aadd08749aef6194370b43f (diff)
Merge branch 'master' into 'master'
add relu layer See merge request !2
Diffstat (limited to 'nerv/matrix')
-rw-r--r--nerv/matrix/generic/cumatrix.c2
-rw-r--r--nerv/matrix/generic/matrix.c23
-rw-r--r--nerv/matrix/generic/mmatrix.c2
3 files changed, 27 insertions, 0 deletions
diff --git a/nerv/matrix/generic/cumatrix.c b/nerv/matrix/generic/cumatrix.c
index 540afd2..3481ede 100644
--- a/nerv/matrix/generic/cumatrix.c
+++ b/nerv/matrix/generic/cumatrix.c
@@ -230,6 +230,8 @@ static const luaL_Reg nerv_matrix_(extra_methods)[] = {
{"sigmoid_grad", nerv_matrix_(lua_sigmoid_grad)},
{"tanh", nerv_matrix_(lua_tanh)},
{"tanh_grad", nerv_matrix_(lua_tanh_grad)},
+ {"relu", nerv_matrix_(lua_relu)},
+ {"relu_grad", nerv_matrix_(lua_relu_grad)},
{"rand_uniform", nerv_matrix_(lua_rand_uniform)},
{"softmax", nerv_matrix_(lua_softmax)},
{"mul_elem", nerv_matrix_(lua_mul_elem)},
diff --git a/nerv/matrix/generic/matrix.c b/nerv/matrix/generic/matrix.c
index 800408d..9f31b4b 100644
--- a/nerv/matrix/generic/matrix.c
+++ b/nerv/matrix/generic/matrix.c
@@ -430,4 +430,27 @@ static int nerv_matrix_(lua_tanh_grad)(lua_State *L) {
return 0;
}
+static int nerv_matrix_(lua_relu)(lua_State *L) {
+ Status status;
+ MATRIX_CONTEXT *context;
+ MATRIX_GET_CONTEXT(L, 3);
+ Matrix *a = luaT_checkudata(L, 1, nerv_matrix_(tname));
+ Matrix *b = luaT_checkudata(L, 2, nerv_matrix_(tname));
+ nerv_matrix_(relu)(a, b, context, &status);
+ NERV_LUA_CHECK_STATUS(L, status);
+ return 0;
+}
+
+static int nerv_matrix_(lua_relu_grad)(lua_State *L) {
+ Status status;
+ MATRIX_CONTEXT *context;
+ MATRIX_GET_CONTEXT(L, 4);
+ Matrix *nerr = luaT_checkudata(L, 1, nerv_matrix_(tname));
+ Matrix *err = luaT_checkudata(L, 2, nerv_matrix_(tname));
+ Matrix *output = luaT_checkudata(L, 3, nerv_matrix_(tname));
+ nerv_matrix_(relu_grad)(nerr, err, output, context, &status);
+ NERV_LUA_CHECK_STATUS(L, status);
+ return 0;
+}
+
#endif
diff --git a/nerv/matrix/generic/mmatrix.c b/nerv/matrix/generic/mmatrix.c
index c03aee4..530888b 100644
--- a/nerv/matrix/generic/mmatrix.c
+++ b/nerv/matrix/generic/mmatrix.c
@@ -122,6 +122,8 @@ static const luaL_Reg nerv_matrix_(extra_methods)[] = {
{"sigmoid_grad", nerv_matrix_(lua_sigmoid_grad)},
{"tanh", nerv_matrix_(lua_tanh)},
{"tanh_grad", nerv_matrix_(lua_tanh_grad)},
+ {"relu", nerv_matrix_(lua_relu)},
+ {"relu_grad", nerv_matrix_(lua_relu_grad)},
{"softmax", nerv_matrix_(lua_softmax)},
{"mul_elem", nerv_matrix_(lua_mul_elem)},
{"log_elem", nerv_matrix_(lua_log_elem)},