summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--embedding_example/setup_nerv.lua10
-rw-r--r--embedding_example/swb_baseline_decode.lua5
m---------speech0
3 files changed, 11 insertions, 4 deletions
diff --git a/embedding_example/setup_nerv.lua b/embedding_example/setup_nerv.lua
index 3ae878d..49a5dd6 100644
--- a/embedding_example/setup_nerv.lua
+++ b/embedding_example/setup_nerv.lua
@@ -7,17 +7,19 @@ param_repo:import(gconf.initialized_param, nil, gconf)
local sublayer_repo = make_sublayer_repo(param_repo)
local layer_repo = make_layer_repo(sublayer_repo, param_repo)
local network = get_network(layer_repo)
+local global_transf = get_global_transf(layer_repo)
local batch_size = 1
network:init(batch_size)
function propagator(input, output)
- local gpu_input = nerv.CuMatrixFloat(input:nrow(), input:ncol())
+ local transformed = nerv.speech_utils.global_transf(input,
+ global_transf, 0, gconf) -- preprocessing
+ local gpu_input = nerv.CuMatrixFloat(transformed:nrow(), transformed:ncol())
local gpu_output = nerv.CuMatrixFloat(output:nrow(), output:ncol())
- gpu_input:copy_fromh(input)
- print(gpu_input)
+ print(transformed)
+ gpu_input:copy_fromh(transformed)
network:propagate({gpu_input}, {gpu_output})
gpu_output:copy_toh(output)
- print(output)
-- collect garbage in-time to save GPU memory
collectgarbage("collect")
end
diff --git a/embedding_example/swb_baseline_decode.lua b/embedding_example/swb_baseline_decode.lua
index 14a463b..8cdb320 100644
--- a/embedding_example/swb_baseline_decode.lua
+++ b/embedding_example/swb_baseline_decode.lua
@@ -107,3 +107,8 @@ end
function get_network(layer_repo)
return layer_repo:get_layer("main")
end
+
+
+function get_global_transf(layer_repo)
+ return layer_repo:get_layer("global_transf")
+end
diff --git a/speech b/speech
-Subproject 261aed43b863c85ecf709dcac2739c388ec491d
+Subproject c6c644223dc9168befd189bf6f33243390671c9