diff options
author | txh18 <[email protected]> | 2015-12-06 13:33:26 +0800 |
---|---|---|
committer | txh18 <[email protected]> | 2015-12-06 13:33:26 +0800 |
commit | 79c711d9c92a8e92f7ad9187a66d3e2aac239356 (patch) | |
tree | 4310e2dcd54f735780a92d43d7c485ca37abfdad | |
parent | ea0e37892ae70357305da3b1fbae617215a25778 (diff) |
small bug fix in lm training script
-rw-r--r-- | nerv/examples/lmptb/lm_trainer.lua | 8 | ||||
-rw-r--r-- | nerv/examples/lmptb/lstmlm_ptb_main.lua | 10 |
2 files changed, 10 insertions, 8 deletions
diff --git a/nerv/examples/lmptb/lm_trainer.lua b/nerv/examples/lmptb/lm_trainer.lua index 6bd06bb..2cdbd4f 100644 --- a/nerv/examples/lmptb/lm_trainer.lua +++ b/nerv/examples/lmptb/lm_trainer.lua @@ -22,7 +22,7 @@ function LMTrainer.lm_process_file_rnn(global_conf, fn, tnn, do_train, p_conf) p_conf = {} end local reader - local r_conf + local r_conf = {} local chunk_size, batch_size if p_conf.one_sen_report == true then --report log prob one by one sentence if do_train == true then @@ -48,6 +48,7 @@ function LMTrainer.lm_process_file_rnn(global_conf, fn, tnn, do_train, p_conf) end global_conf.timer:flush() + tnn:init(batch_size, chunk_size) tnn:flush_all() --caution: will also flush the inputs from the reader! local next_log_wcn = global_conf.log_w_num @@ -107,7 +108,7 @@ function LMTrainer.lm_process_file_rnn(global_conf, fn, tnn, do_train, p_conf) end if p_conf.one_sen_report == true then for i = 1, batch_size do - nerv.printf("LMTrainer.lm_process_file_rnn: one_sen_report, %f\n", sen_logp[i]) + nerv.printf("LMTrainer.lm_process_file_rnn: one_sen_report_output, %f\n", sen_logp[i]) end end @@ -177,6 +178,7 @@ function LMTrainer.lm_process_file_birnn(global_conf, fn, tnn, do_train, p_conf) end global_conf.timer:flush() + tnn:init(batch_size, chunk_size) tnn:flush_all() --caution: will also flush the inputs from the reader! local next_log_wcn = global_conf.log_w_num @@ -235,7 +237,7 @@ function LMTrainer.lm_process_file_birnn(global_conf, fn, tnn, do_train, p_conf) end if p_conf.one_sen_report == true then for i = 1, batch_size do - nerv.printf("LMTrainer.lm_process_file_birnn: one_sen_report, %f\n", sen_logp[i]) + nerv.printf("LMTrainer.lm_process_file_birnn: one_sen_report_output, %f\n", sen_logp[i]) end end diff --git a/nerv/examples/lmptb/lstmlm_ptb_main.lua b/nerv/examples/lmptb/lstmlm_ptb_main.lua index 2751ec8..804438d 100644 --- a/nerv/examples/lmptb/lstmlm_ptb_main.lua +++ b/nerv/examples/lmptb/lstmlm_ptb_main.lua @@ -337,9 +337,8 @@ ppl_last = 100000 commands_str = "train:test" commands = {} test_iter = -1 - --for testout(question) -local q_file = "ptb.test.txt.q10rs1_Msss.adds" +q_file = "ptb.test.txt.q10rs1_Msss.adds" if arg[2] ~= nil then nerv.printf("%s applying arg[2](%s)...\n", global_conf.sche_log_pre, arg[2]) @@ -485,8 +484,8 @@ end --if commands["test"] if commands["testout"] == 1 then nerv.printf("===TEST OUT===\n") nerv.printf("q_file:\t%s\n", q_file) - local q_fn = qdata_dir .. q_file - global_conf.sche_log_pre = "[SCHEDULER FINAL_TEST]:" + local q_fn = qdata_dir .. '/' .. q_file + global_conf.sche_log_pre = "[SCHEDULER TESTOUT]:" if final_iter ~= -1 and test_iter == -1 then test_iter = final_iter end @@ -495,7 +494,8 @@ if commands["testout"] == 1 then end tnn = load_net(global_conf, test_iter) global_conf.dropout_rate = 0 - LMTrainer.lm_process_file_rnn(global_conf, q_fn, tnn, false) --false update! + LMTrainer.lm_process_file_rnn(global_conf, q_fn, tnn, false, + {["one_sen_report"] = true}) --false update! end --if commands["testout"] |