From 126433c7ab118e6ca65a3567662cec9509075cef Mon Sep 17 00:00:00 2001 From: Determinant Date: Wed, 12 Aug 2015 15:30:25 +0800 Subject: add script to convert nerv param file to TNet format --- htk_io/tools/nerv_to_tnet.lua | 60 +++++++++++++++++++++++++++++++++++++++++++ htk_io/tools/tnet_to_nerv.c | 2 +- 2 files changed, 61 insertions(+), 1 deletion(-) create mode 100644 htk_io/tools/nerv_to_tnet.lua (limited to 'htk_io') diff --git a/htk_io/tools/nerv_to_tnet.lua b/htk_io/tools/nerv_to_tnet.lua new file mode 100644 index 0000000..c0ac76b --- /dev/null +++ b/htk_io/tools/nerv_to_tnet.lua @@ -0,0 +1,60 @@ +-- usage: nerv config_file nerv_param_input tnet_output + +dofile(arg[1]) +param_repo = nerv.ParamRepo() +param_repo:import({arg[2], gconf.initialized_param[2]}, nil, gconf) +layer_repo = make_layer_repo(param_repo) +f = assert(io.open(arg[3], "w")) + +function print_tnet_matrix(cumat) + local strs = {} + collectgarbage() + if cumat:nrow() == 1 then + local mat = nerv.MMatrixFloat(1, cumat:ncol()) + cumat:copy_toh(mat) + table.insert(strs, string.format("v %d\n", mat:ncol())) + for j = 0, mat:ncol() - 1 do + table.insert(strs, string.format("%.8f ", mat[0][j])) + end + table.insert(strs, "\n") + f:write(table.concat(strs)) + else + cumat = cumat:trans() + local mat = nerv.MMatrixFloat(cumat:nrow(), cumat:ncol()) + cumat:copy_toh(mat) + table.insert(strs, string.format("m %d %d\n", mat:nrow(), mat:ncol())) + for i = 0, mat:nrow() - 1 do + local row = mat[i] + for j = 0, mat:ncol() - 1 do + table.insert(strs, string.format("%.8f ", row[j])) + end + table.insert(strs, "\n") + f:write(table.concat(strs)) + strs = {} + end + end +end +local lnames = {"affine0", "sigmoid0", + "affine1", "sigmoid1", + "affine2", "sigmoid2", + "affine3", "sigmoid3", + "affine4", "sigmoid4", + "affine5", "sigmoid5", + "affine6", "sigmoid6", + "affine7", "ce_crit"} +for i, name in ipairs(lnames) do + local layer = layer_repo:get_layer(name) + local layer_type = layer.__typename + if layer_type == "nerv.AffineLayer" then + f:write(string.format(" %d %d\n", layer.dim_out[1], layer.dim_in[1])) + print_tnet_matrix(layer.ltp.trans) + print_tnet_matrix(layer.bp.trans) + elseif layer_type == "nerv.SigmoidLayer" then + f:write(string.format(" %d %d\n", layer.dim_out[1], layer.dim_in[1])) + elseif layer_type == "nerv.SoftmaxCELayer" then + f:write(string.format(" %d %d\n", layer.dim_in[1], layer.dim_in[1])) + else + nerv.error("unknown layer type %s", layer_type) + end +end +f:close() diff --git a/htk_io/tools/tnet_to_nerv.c b/htk_io/tools/tnet_to_nerv.c index f781236..5774819 100644 --- a/htk_io/tools/tnet_to_nerv.c +++ b/htk_io/tools/tnet_to_nerv.c @@ -41,7 +41,7 @@ int main() { fprintf(fout, "%16d", 0); fprintf(fout, "{type=\"nerv.BiasParam\",id=\"affine%d_bp\"}\n", cnt); - fprintf(fout, "1 %d\n", nrow, ncol); + fprintf(fout, "1 %d\n", ncol); for (j = 0; j < ncol; j++) fprintf(fout, "%.8f ", mat[0][j]); fprintf(fout, "\n"); -- cgit v1.2.3-70-g09d2 From 184a2710192950fb62269726243a240ff18294eb Mon Sep 17 00:00:00 2001 From: Determinant Date: Thu, 13 Aug 2015 15:36:09 +0800 Subject: [tnet_to_nerv] add support for matrix of arbitrary size --- htk_io/tools/tnet_to_nerv.cpp | 8 +++++++- 1 file changed, 7 insertions(+), 1 deletion(-) (limited to 'htk_io') diff --git a/htk_io/tools/tnet_to_nerv.cpp b/htk_io/tools/tnet_to_nerv.cpp index bbfddcf..a779a25 100644 --- a/htk_io/tools/tnet_to_nerv.cpp +++ b/htk_io/tools/tnet_to_nerv.cpp @@ -2,9 +2,10 @@ #include #include #include +#include char token[1024]; char output[1024]; -double mat[4096][4096]; +double **mat; int main(int argc, char **argv) { std::ofstream fout; fout.open(argv[1]); @@ -18,6 +19,9 @@ int main(int argc, char **argv) { scanf("%d %d", &ncol, &nrow); scanf("%s %d %d", token, &ncol, &nrow); printf("%d %d\n", nrow, ncol); + mat = (double **)malloc(nrow * sizeof(double *)); + for (i = 0; i < nrow; i++) + mat[i] = (double *)malloc(ncol * sizeof(double)); for (j = 0; j < ncol; j++) for (i = 0; i < nrow; i++) scanf("%lf", mat[i] + j); @@ -34,7 +38,9 @@ int main(int argc, char **argv) { for (j = 0; j < ncol; j++) fout << mat[i][j] << " "; fout << std::endl; + free(mat[i]); } + free(mat); long length = fout.tellp() - base; fout.seekp(base); sprintf(output, "[%13lu]\n", length); -- cgit v1.2.3-70-g09d2 From c177a7549bd90670af4b29fa813ddea32cfe0f78 Mon Sep 17 00:00:00 2001 From: Determinant Date: Thu, 13 Aug 2015 15:57:58 +0800 Subject: add frm_trim option to global_transf --- htk_io/init.lua | 3 ++- speech_utils/init.lua | 7 ++++--- 2 files changed, 6 insertions(+), 4 deletions(-) (limited to 'htk_io') diff --git a/htk_io/init.lua b/htk_io/init.lua index af92140..c4dfff9 100644 --- a/htk_io/init.lua +++ b/htk_io/init.lua @@ -34,7 +34,8 @@ function TNetReader:get_data() -- global transf if self.global_transf ~= nil then feat_utter = nerv.speech_utils.global_transf(feat_utter, - self.global_transf, self.frm_ext, self.gconf) + self.global_transf, self.frm_ext, + self.frm_ext, self.gconf) end res[self.feat_id] = feat_utter -- add corresponding labels diff --git a/speech_utils/init.lua b/speech_utils/init.lua index 3d1a470..9690e71 100644 --- a/speech_utils/init.lua +++ b/speech_utils/init.lua @@ -1,6 +1,7 @@ nerv.speech_utils = {} -function nerv.speech_utils.global_transf(feat_utter, global_transf, frm_ext, gconf) +function nerv.speech_utils.global_transf(feat_utter, global_transf, + frm_ext, frm_trim, gconf) local rearranged feat_utter = gconf.cumat_type.new_from_host(feat_utter) if frm_ext > 0 then @@ -21,8 +22,8 @@ function nerv.speech_utils.global_transf(feat_utter, global_transf, frm_ext, gco global_transf:init(input[1]:nrow()) global_transf:propagate(input, output) -- trim frames - expanded = gconf.mmat_type(output[1]:nrow() - frm_ext * 2, output[1]:ncol()) - output[1]:copy_toh(expanded, frm_ext, feat_utter:nrow() - frm_ext) + expanded = gconf.mmat_type(output[1]:nrow() - frm_trim * 2, output[1]:ncol()) + output[1]:copy_toh(expanded, frm_trim, feat_utter:nrow() - frm_trim) collectgarbage("collect") return expanded end -- cgit v1.2.3-70-g09d2