summaryrefslogtreecommitdiff
path: root/examples
diff options
context:
space:
mode:
authorDeterminant <[email protected]>2015-06-03 23:00:30 +0800
committerDeterminant <[email protected]>2015-06-03 23:00:30 +0800
commita753eca0121ac3ec81ed76bd719d3f1cb9522680 (patch)
tree9777fdddf5d0404964353a0b3d2821e514f6eeb3 /examples
parent38962683e518dcbebc0cfa6c0c9c9616b25d5bd1 (diff)
...
Diffstat (limited to 'examples')
-rw-r--r--examples/tnet_preprocessing_example.lua2
-rw-r--r--examples/tnet_preprocessing_example2.lua2
-rw-r--r--examples/tnet_sgd_buffer.lua70
3 files changed, 72 insertions, 2 deletions
diff --git a/examples/tnet_preprocessing_example.lua b/examples/tnet_preprocessing_example.lua
index 8a65b44..5f20910 100644
--- a/examples/tnet_preprocessing_example.lua
+++ b/examples/tnet_preprocessing_example.lua
@@ -1,6 +1,6 @@
require 'libspeech'
frm_ext = 5
-gconf = {mat_type = nerv.CuMatrixFloat,
+gconf = {cumat_type = nerv.CuMatrixFloat,
batch_size = 158}
param_repo = nerv.ParamRepo({"global_transf.nerv"})
sublayer_repo = nerv.LayerRepo(
diff --git a/examples/tnet_preprocessing_example2.lua b/examples/tnet_preprocessing_example2.lua
index ae8d86b..8e1bc85 100644
--- a/examples/tnet_preprocessing_example2.lua
+++ b/examples/tnet_preprocessing_example2.lua
@@ -1,5 +1,5 @@
require 'speech.init'
-gconf = {mat_type = nerv.CuMatrixFloat,
+gconf = {cumat_type = nerv.CuMatrixFloat,
batch_size = 158}
param_repo = nerv.ParamRepo({"global_transf.nerv"})
diff --git a/examples/tnet_sgd_buffer.lua b/examples/tnet_sgd_buffer.lua
new file mode 100644
index 0000000..152d2f5
--- /dev/null
+++ b/examples/tnet_sgd_buffer.lua
@@ -0,0 +1,70 @@
+require 'speech.init'
+gconf = {cumat_type = nerv.CuMatrixFloat,
+ mmat_type = nerv.MMatrixFloat,
+ batch_size = 256}
+param_repo = nerv.ParamRepo({"global_transf.nerv"})
+
+sublayer_repo = nerv.LayerRepo(
+ {
+ ["nerv.BiasLayer"] =
+ {
+ blayer1 = {{bias = "bias1"}, {dim_in = {429}, dim_out = {429}}},
+ blayer2 = {{bias = "bias2"}, {dim_in = {429}, dim_out = {429}}}
+ },
+ ["nerv.WindowLayer"] =
+ {
+ wlayer1 = {{window = "window1"}, {dim_in = {429}, dim_out = {429}}},
+ wlayer2 = {{window = "window2"}, {dim_in = {429}, dim_out = {429}}}
+ }
+ }, param_repo, gconf)
+
+layer_repo = nerv.LayerRepo(
+ {
+ ["nerv.DAGLayer"] =
+ {
+ main = {{}, {
+ dim_in = {429}, dim_out = {429},
+ sub_layers = sublayer_repo,
+ connections = {
+ ["<input>[1]"] = "blayer1[1]",
+ ["blayer1[1]"] = "wlayer1[1]",
+ ["wlayer1[1]"] = "blayer2[1]",
+ ["blayer2[1]"] = "wlayer2[1]",
+ ["wlayer2[1]"] = "<output>[1]"
+ }
+ }}
+ }
+ }, param_repo, gconf)
+
+tnet_reader = nerv.TNetReader({},
+ {
+ id = "main_scp",
+-- scp_file = "/slfs1/users/mfy43/swb_ivec/train_bp.scp",
+ scp_file = "t.scp",
+ conf_file = "/slfs1/users/mfy43/swb_ivec/plp_0_d_a.conf",
+ frm_ext = 5,
+ mlfs = {
+ ref = {
+ file = "/slfs1/users/mfy43/swb_ivec/ref.mlf",
+ format = "map",
+ format_arg = "/slfs1/users/mfy43/swb_ivec/dict",
+ dir = "*/",
+ ext = "lab"
+ }
+ },
+ global_transf = layer_repo:get_layer("main")
+ })
+
+buffer = nerv.SGDBuffer(gconf,
+ {
+ buffer_size = 1024,
+ readers = {
+ { reader = tnet_reader,
+ data = {main_scp = 429, ref = 1}}
+ }
+ })
+
+for data in buffer.get_data, buffer do
+ print(data.main_scp)
+-- print(data.ref)
+end