diff options
Diffstat (limited to 'nerv/examples/lmptb/main.lua')
-rw-r--r-- | nerv/examples/lmptb/main.lua | 11 |
1 files changed, 6 insertions, 5 deletions
diff --git a/nerv/examples/lmptb/main.lua b/nerv/examples/lmptb/main.lua index 9b39e83..13d610e 100644 --- a/nerv/examples/lmptb/main.lua +++ b/nerv/examples/lmptb/main.lua @@ -15,7 +15,7 @@ function prepare_parameters(global_conf, first_time) if (first_time) then ltp_ih = nerv.LinearTransParam("ltp_ih", global_conf) - ltp_ih.trans = global_conf.cumat_type(global_conf.vocab:size(), global_conf.hidden_size) + ltp_ih.trans = global_conf.cumat_type(global_conf.vocab:size() + 1, global_conf.hidden_size) --index 0 is for zero, others correspond to vocab index(starting from 1) ltp_ih.trans:generate(global_conf.param_random) ltp_hh = nerv.LinearTransParam("ltp_hh", global_conf) @@ -164,7 +164,7 @@ function propagateFile(global_conf, dagL, fn, config) local dagL_input = {} for i = 1, global_conf.bptt + 1 do - dagL_input[i] = nerv.MMatrixInt(global_conf.batch_size, 1) + dagL_input[i] = global_conf.cumat_type(1, global_conf.batch_size) --changed to row vector, debughtx end dagL_input[global_conf.bptt + 2] = global_conf.cumat_type(global_conf.batch_size, global_conf.hidden_size) dagL_input[global_conf.bptt + 3] = global_conf.cumat_type(global_conf.batch_size, global_conf.vocab:size()) @@ -192,7 +192,7 @@ function propagateFile(global_conf, dagL, fn, config) global_conf.input_word_id["recurrentL"..i] = dagL_input[i] --for IndRecurrent end dagL_input[global_conf.bptt + 2]:copy_fromd(hidden_store[tnow - global_conf.bptt - 1]) - nerv.LMUtil.set_onehot(dagL_input[global_conf.bptt + 3], token_store[tnow + 1], global_conf.vocab) + nerv.LMUtil.set_onehot(dagL_input[global_conf.bptt + 3], token_store[tnow + 1], global_conf.vocab) --for softmax --local dagL_input = create_dag_input(global_conf, token_store, hidden_store, tnow) global_conf.timer:tic("dagL-propagate") @@ -224,6 +224,7 @@ function propagateFile(global_conf, dagL, fn, config) for key, value in pairs(global_conf.timer.rec) do printf("\t [global_conf.timer]: time spent on %s:%.5fs\n", key, value) end + --comment this for debughtx global_conf.timer:flush() --nerv.CuMatrix.print_profile() --nerv.CuMatrix.clear_profile() @@ -277,10 +278,10 @@ if (set == "ptb") then valid_fn = valid_fn, test_fn = test_fn, sche_log_pre = "[SCHEDULER]:", - log_w_num = 10000, --give a message when log_w_num words have been processed + log_w_num = 500000, --give a message when log_w_num words have been processed timer = nerv.Timer() } - global_conf.work_dir = work_dir_base.."/h"..global_conf.hidden_size.."bp"..global_conf.bptt.."slr"..global_conf.lrate..os.date("_%bD%dH%H") + global_conf.work_dir = work_dir_base.."/h"..global_conf.hidden_size.."bp"..global_conf.bptt.."slr"..global_conf.lrate --..os.date("_%bD%dH%H") --comment this for testing global_conf.param_fn = global_conf.work_dir.."/params" elseif (set == "test") then train_fn = "/slfs1/users/txh18/workspace/nerv-project/some-text" |