From a74183ddb4ab8383bfe214b3745eb8a0a99ee47a Mon Sep 17 00:00:00 2001 From: Determinant Date: Thu, 25 Jun 2015 12:56:45 +0800 Subject: let HTK I/O implementation be a single package --- examples/tnet_io_example.lua | 16 ------- examples/tnet_preprocessing_example.lua | 75 -------------------------------- examples/tnet_preprocessing_example2.lua | 68 ----------------------------- examples/tnet_sgd_buffer.lua | 70 ----------------------------- 4 files changed, 229 deletions(-) delete mode 100644 examples/tnet_io_example.lua delete mode 100644 examples/tnet_preprocessing_example.lua delete mode 100644 examples/tnet_preprocessing_example2.lua delete mode 100644 examples/tnet_sgd_buffer.lua (limited to 'examples') diff --git a/examples/tnet_io_example.lua b/examples/tnet_io_example.lua deleted file mode 100644 index eea73a5..0000000 --- a/examples/tnet_io_example.lua +++ /dev/null @@ -1,16 +0,0 @@ -require 'libspeech' -frm_ext = 5 -feat_repo = nerv.TNetFeatureRepo( - "/slfs1/users/mfy43/swb_ivec/train_bp.scp", - "/slfs1/users/mfy43/swb_ivec/plp_0_d_a.conf", - frm_ext) -lab_repo = nerv.TNetLabelRepo( - "/slfs1/users/mfy43/swb_ivec/ref.mlf", - "map", - "/slfs1/users/mfy43/swb_ivec/dict", - "*/", - "lab") -feat_utter = feat_repo:cur_utter() -print(feat_utter) -lab_utter = lab_repo:get_utter(feat_repo, feat_utter:nrow() - frm_ext * 2) -print(lab_utter) diff --git a/examples/tnet_preprocessing_example.lua b/examples/tnet_preprocessing_example.lua deleted file mode 100644 index 9e1c0ce..0000000 --- a/examples/tnet_preprocessing_example.lua +++ /dev/null @@ -1,75 +0,0 @@ -require 'libspeech' -frm_ext = 5 -gconf = {cumat_type = nerv.CuMatrixFloat, - batch_size = 158} -param_repo = nerv.ParamRepo({"global_transf.nerv"}) -sublayer_repo = nerv.LayerRepo( - { - ["nerv.BiasLayer"] = - { - blayer1 = {{bias = "bias1"}, {dim_in = {429}, dim_out = {429}}}, - blayer2 = {{bias = "bias2"}, {dim_in = {429}, dim_out = {429}}} - }, - ["nerv.WindowLayer"] = - { - wlayer1 = {{window = "window1"}, {dim_in = {429}, dim_out = {429}}}, - wlayer2 = {{window = "window2"}, {dim_in = {429}, dim_out = {429}}} - } - }, param_repo, gconf) - -layer_repo = nerv.LayerRepo( - { - ["nerv.DAGLayer"] = - { - main = {{}, { - dim_in = {429}, dim_out = {429}, - sub_layers = sublayer_repo, - connections = { - ["[1]"] = "blayer1[1]", - ["blayer1[1]"] = "wlayer1[1]", - ["wlayer1[1]"] = "blayer2[1]", - ["blayer2[1]"] = "wlayer2[1]", - ["wlayer2[1]"] = "[1]" - } - }} - } - }, param_repo, gconf) - -feat_repo = nerv.TNetFeatureRepo( - "/slfs1/users/mfy43/swb_ivec/train_bp.scp", - "/slfs1/users/mfy43/swb_ivec/plp_0_d_a.conf", - frm_ext) -lab_repo = nerv.TNetLabelRepo( - "/slfs1/users/mfy43/swb_ivec/ref.mlf", - "map", - "/slfs1/users/mfy43/swb_ivec/dict", - "*/", - "lab") -feat_utter = feat_repo:cur_utter() - --- print(feat_utter) --- lab_utter = lab_repo:get_utter(feat_repo, feat_utter:nrow() - frm_ext * 2) --- print(lab_utter) - -cf2 = nerv.ChunkFile("feat_256", "r") -input = cf2:read_chunk("input", gconf) - -step = frm_ext * 2 + 1 -expanded = nerv.CuMatrixFloat(feat_utter:nrow(), feat_utter:ncol() * step) -expanded:expand_frm(nerv.CuMatrixFloat.new_from_host(feat_utter), frm_ext) - -rearranged = expanded:create() -rearranged:rearrange_frm(expanded, step) - -output = {expanded:create()} -main = layer_repo:get_layer("main") -main:init() -main:propagate({rearranged}, output) - -for i = 0, 157 - 10 do - row_diff = input.trans[i] - output[1][i + 5] - for j = 0, row_diff:ncol() - 1 do - nerv.printf("%.8f ", row_diff[j]) - end - nerv.printf("\n") -end diff --git a/examples/tnet_preprocessing_example2.lua b/examples/tnet_preprocessing_example2.lua deleted file mode 100644 index 1215b23..0000000 --- a/examples/tnet_preprocessing_example2.lua +++ /dev/null @@ -1,68 +0,0 @@ -require 'speech.init' -gconf = {cumat_type = nerv.CuMatrixFloat, - batch_size = 158} -param_repo = nerv.ParamRepo({"global_transf.nerv"}) - -sublayer_repo = nerv.LayerRepo( - { - ["nerv.BiasLayer"] = - { - blayer1 = {{bias = "bias1"}, {dim_in = {429}, dim_out = {429}}}, - blayer2 = {{bias = "bias2"}, {dim_in = {429}, dim_out = {429}}} - }, - ["nerv.WindowLayer"] = - { - wlayer1 = {{window = "window1"}, {dim_in = {429}, dim_out = {429}}}, - wlayer2 = {{window = "window2"}, {dim_in = {429}, dim_out = {429}}} - } - }, param_repo, gconf) - -layer_repo = nerv.LayerRepo( - { - ["nerv.DAGLayer"] = - { - main = {{}, { - dim_in = {429}, dim_out = {429}, - sub_layers = sublayer_repo, - connections = { - ["[1]"] = "blayer1[1]", - ["blayer1[1]"] = "wlayer1[1]", - ["wlayer1[1]"] = "blayer2[1]", - ["blayer2[1]"] = "wlayer2[1]", - ["wlayer2[1]"] = "[1]" - } - }} - } - }, param_repo, gconf) - -reader = nerv.TNetReader({}, - { - id = "main_scp", - scp_file = "/slfs1/users/mfy43/swb_ivec/train_bp.scp", - conf_file = "/slfs1/users/mfy43/swb_ivec/plp_0_d_a.conf", - frm_ext = 5, - mlfs = { - ref = { - file = "/slfs1/users/mfy43/swb_ivec/ref.mlf", - format = "map", - format_arg = "/slfs1/users/mfy43/swb_ivec/dict", - dir = "*/", - ext = "lab" - } - }, - global_transf = layer_repo:get_layer("main") - }) - -utter = reader:get_data() --- print(utter.main_scp) -print(utter.ref) --- cf2 = nerv.ChunkFile("feat_256", "r") --- input = cf2:read_chunk("input", gconf) - --- for i = 0, 157 - 10 do --- row_diff = input.trans[i] - utter.main_scp[i] --- for j = 0, row_diff:ncol() - 1 do --- nerv.printf("%.8f ", row_diff[j]) --- end --- nerv.printf("\n") --- end diff --git a/examples/tnet_sgd_buffer.lua b/examples/tnet_sgd_buffer.lua deleted file mode 100644 index 152d2f5..0000000 --- a/examples/tnet_sgd_buffer.lua +++ /dev/null @@ -1,70 +0,0 @@ -require 'speech.init' -gconf = {cumat_type = nerv.CuMatrixFloat, - mmat_type = nerv.MMatrixFloat, - batch_size = 256} -param_repo = nerv.ParamRepo({"global_transf.nerv"}) - -sublayer_repo = nerv.LayerRepo( - { - ["nerv.BiasLayer"] = - { - blayer1 = {{bias = "bias1"}, {dim_in = {429}, dim_out = {429}}}, - blayer2 = {{bias = "bias2"}, {dim_in = {429}, dim_out = {429}}} - }, - ["nerv.WindowLayer"] = - { - wlayer1 = {{window = "window1"}, {dim_in = {429}, dim_out = {429}}}, - wlayer2 = {{window = "window2"}, {dim_in = {429}, dim_out = {429}}} - } - }, param_repo, gconf) - -layer_repo = nerv.LayerRepo( - { - ["nerv.DAGLayer"] = - { - main = {{}, { - dim_in = {429}, dim_out = {429}, - sub_layers = sublayer_repo, - connections = { - ["[1]"] = "blayer1[1]", - ["blayer1[1]"] = "wlayer1[1]", - ["wlayer1[1]"] = "blayer2[1]", - ["blayer2[1]"] = "wlayer2[1]", - ["wlayer2[1]"] = "[1]" - } - }} - } - }, param_repo, gconf) - -tnet_reader = nerv.TNetReader({}, - { - id = "main_scp", --- scp_file = "/slfs1/users/mfy43/swb_ivec/train_bp.scp", - scp_file = "t.scp", - conf_file = "/slfs1/users/mfy43/swb_ivec/plp_0_d_a.conf", - frm_ext = 5, - mlfs = { - ref = { - file = "/slfs1/users/mfy43/swb_ivec/ref.mlf", - format = "map", - format_arg = "/slfs1/users/mfy43/swb_ivec/dict", - dir = "*/", - ext = "lab" - } - }, - global_transf = layer_repo:get_layer("main") - }) - -buffer = nerv.SGDBuffer(gconf, - { - buffer_size = 1024, - readers = { - { reader = tnet_reader, - data = {main_scp = 429, ref = 1}} - } - }) - -for data in buffer.get_data, buffer do - print(data.main_scp) --- print(data.ref) -end -- cgit v1.2.3