diff options
author | TianxingHe <[email protected]> | 2016-01-19 20:13:49 -0800 |
---|---|---|
committer | TianxingHe <[email protected]> | 2016-01-19 20:13:49 -0800 |
commit | dcad8a3f80fc55ca93984d981f9b829d2e4ea728 (patch) | |
tree | 61b9bc1d043883bb5d85dcb86cfb621396d75c41 /nerv/matrix/generic | |
parent | 7449dd19c4d1669b483693f61add9d574e46f0b2 (diff) | |
parent | 37dec2610c92d03813c4e91ed58791ab60da6646 (diff) |
Merge pull request #21 from cloudygoose/txh18/rnnlm
Txh18/rnnlm new changes to lm side
Diffstat (limited to 'nerv/matrix/generic')
-rw-r--r-- | nerv/matrix/generic/cumatrix.c | 36 | ||||
-rw-r--r-- | nerv/matrix/generic/elem_type.h | 22 | ||||
-rw-r--r-- | nerv/matrix/generic/mmatrix.c | 2 |
3 files changed, 33 insertions, 27 deletions
diff --git a/nerv/matrix/generic/cumatrix.c b/nerv/matrix/generic/cumatrix.c index fb36033..edd7b0a 100644 --- a/nerv/matrix/generic/cumatrix.c +++ b/nerv/matrix/generic/cumatrix.c @@ -1,5 +1,5 @@ #ifdef NERV_GENERIC_CUMATRIX -#include "elem_type.h" +#include "../../lib/matrix/generic/elem_type.h" #define MATRIX_DATA_WRITE(L, data, idx, val) cuda_matrix_(write)(L, data, idx, val) #define MATRIX_DATA_READ(L, data, idx) cuda_matrix_(read)(L, data, idx) #define MATRIX_INIT(L) cuda_matrix_(init)(L) @@ -292,6 +292,18 @@ static int nerv_matrix_(lua_copy_rows_fromd_by_idx)(lua_State *L) { return 0; } +static int nerv_matrix_(lua_copy_rows_fromd_by_colidx)(lua_State *L) { + Status status; + Matrix *a = luaT_checkudata(L, 1, nerv_matrix_(tname)); + const Matrix *b = luaT_checkudata(L, 2, nerv_matrix_(tname)); + const Matrix *idx = luaT_checkudata(L, 3, nerv_matrix_(tname)); + long nrow = a->nrow; + int idx_begin = lua_gettop(L) > 3 ? luaL_checkinteger(L, 4) : 0; + nerv_matrix_(copy_rows_fromd_by_colidx)(a, b, idx, idx_begin, &status); + NERV_LUA_CHECK_STATUS(L, status); + return 0; +} + static int nerv_matrix_(lua_expand_frm)(lua_State *L) { Status status; Matrix *a = luaT_checkudata(L, 1, nerv_matrix_(tname)); @@ -331,7 +343,21 @@ static int nerv_matrix_(lua_scale_rows_by_row)(lua_State *L) { } #ifdef __NERV_FUTURE_CUDA_7 -static int nerv_matrix_(lua_update_select_rows)(lua_State *L) { +static int nerv_matrix_(lua_update_select_rows_by_rowidx)(lua_State *L) { + /* update c's select rows, + * i.e. c[idx[i]] = c[idx[i]] * (1 - beta * alpha) + a[i] * alpha */ + Status status; + Matrix *c = luaT_checkudata(L, 1, nerv_matrix_(tname)); + const Matrix *a = luaT_checkudata(L, 2, nerv_matrix_(tname)); + const Matrix *idx = luaT_checkudata(L, 3, nerv_matrix_(tname)); + MATRIX_ELEM alpha = luaL_checknumber(L, 4); + MATRIX_ELEM beta = luaL_checknumber(L, 5); + nerv_matrix_(update_select_rows_by_rowidx)(c, a, idx, alpha, beta, &status); + NERV_LUA_CHECK_STATUS(L, status); + return 0; +} + +static int nerv_matrix_(lua_update_select_rows_by_colidx)(lua_State *L) { /* update c's select rows, * i.e. c[idx[i]] = c[idx[i]] * (1 - beta * alpha) + a[i] * alpha */ Status status; @@ -340,7 +366,7 @@ static int nerv_matrix_(lua_update_select_rows)(lua_State *L) { const Matrix *idx = luaT_checkudata(L, 3, nerv_matrix_(tname)); MATRIX_ELEM alpha = luaL_checknumber(L, 4); MATRIX_ELEM beta = luaL_checknumber(L, 5); - nerv_matrix_(update_select_rows)(c, a, idx, alpha, beta, &status); + nerv_matrix_(update_select_rows_by_colidx)(c, a, idx, alpha, beta, &status); NERV_LUA_CHECK_STATUS(L, status); return 0; } @@ -376,12 +402,14 @@ static const luaL_Reg nerv_matrix_(extra_methods)[] = { {"thres_mask", nerv_matrix_(lua_thres_mask)}, {"copy_rows_fromh_by_idx", nerv_matrix_(lua_copy_rows_fromh_by_idx)}, {"copy_rows_fromd_by_idx", nerv_matrix_(lua_copy_rows_fromd_by_idx)}, + {"copy_rows_fromd_by_colidx", nerv_matrix_(lua_copy_rows_fromd_by_colidx)}, {"expand_frm", nerv_matrix_(lua_expand_frm)}, {"rearrange_frm", nerv_matrix_(lua_rearrange_frm)}, {"scale_rows_by_row", nerv_matrix_(lua_scale_rows_by_row)}, {"scale_rows_by_col", nerv_matrix_(lua_scale_rows_by_col)}, #ifdef __NERV_FUTURE_CUDA_7 - {"update_select_rows", nerv_matrix_(lua_update_select_rows)}, + {"update_select_rows_by_rowidx", nerv_matrix_(lua_update_select_rows_by_rowidx)}, + {"update_select_rows_by_colidx", nerv_matrix_(lua_update_select_rows_by_colidx)}, #endif {NULL, NULL} }; diff --git a/nerv/matrix/generic/elem_type.h b/nerv/matrix/generic/elem_type.h deleted file mode 100644 index bffe940..0000000 --- a/nerv/matrix/generic/elem_type.h +++ /dev/null @@ -1,22 +0,0 @@ -#ifdef MATRIX_USE_FLOAT - -#define MATRIX_ELEM float -#define MATRIX_ELEM_FMT "%f" -#define MATRIX_ELEM_WRITE_FMT "%.8f" -#define MATRIX_ELEM_PTR(self) ((self)->data.f) - -#elif defined(MATRIX_USE_DOUBLE) - -#define MATRIX_ELEM double -#define MATRIX_ELEM_FMT "%lf" -#define MATRIX_ELEM_WRITE_FMT "%.8lf" -#define MATRIX_ELEM_PTR(self) ((self)->data.d) - -#elif defined(MATRIX_USE_INT) - -#define MATRIX_ELEM long -#define MATRIX_ELEM_FMT "%ld" -#define MATRIX_ELEM_WRITE_FMT "%ld" -#define MATRIX_ELEM_PTR(self) ((self)->data.i) - -#endif diff --git a/nerv/matrix/generic/mmatrix.c b/nerv/matrix/generic/mmatrix.c index 01dd9e5..a4e8489 100644 --- a/nerv/matrix/generic/mmatrix.c +++ b/nerv/matrix/generic/mmatrix.c @@ -1,6 +1,6 @@ #ifdef NERV_GENERIC_MMATRIX #include "../../lib/matrix/generic/matrix.h" -#include "elem_type.h" +#include "../../lib/matrix/generic/elem_type.h" #define MATRIX_DATA_WRITE(L, data, idx, val) (data[idx] = val) #define MATRIX_DATA_READ(L, data, idx) (data[idx]) #define MATRIX_INIT(L) host_matrix_(init)(L) |