aboutsummaryrefslogtreecommitdiff
path: root/nerv/examples/lmptb/main.lua
diff options
context:
space:
mode:
Diffstat (limited to 'nerv/examples/lmptb/main.lua')
-rw-r--r--nerv/examples/lmptb/main.lua11
1 files changed, 6 insertions, 5 deletions
diff --git a/nerv/examples/lmptb/main.lua b/nerv/examples/lmptb/main.lua
index 9b39e83..13d610e 100644
--- a/nerv/examples/lmptb/main.lua
+++ b/nerv/examples/lmptb/main.lua
@@ -15,7 +15,7 @@ function prepare_parameters(global_conf, first_time)
if (first_time) then
ltp_ih = nerv.LinearTransParam("ltp_ih", global_conf)
- ltp_ih.trans = global_conf.cumat_type(global_conf.vocab:size(), global_conf.hidden_size)
+ ltp_ih.trans = global_conf.cumat_type(global_conf.vocab:size() + 1, global_conf.hidden_size) --index 0 is for zero, others correspond to vocab index(starting from 1)
ltp_ih.trans:generate(global_conf.param_random)
ltp_hh = nerv.LinearTransParam("ltp_hh", global_conf)
@@ -164,7 +164,7 @@ function propagateFile(global_conf, dagL, fn, config)
local dagL_input = {}
for i = 1, global_conf.bptt + 1 do
- dagL_input[i] = nerv.MMatrixInt(global_conf.batch_size, 1)
+ dagL_input[i] = global_conf.cumat_type(1, global_conf.batch_size) --changed to row vector, debughtx
end
dagL_input[global_conf.bptt + 2] = global_conf.cumat_type(global_conf.batch_size, global_conf.hidden_size)
dagL_input[global_conf.bptt + 3] = global_conf.cumat_type(global_conf.batch_size, global_conf.vocab:size())
@@ -192,7 +192,7 @@ function propagateFile(global_conf, dagL, fn, config)
global_conf.input_word_id["recurrentL"..i] = dagL_input[i] --for IndRecurrent
end
dagL_input[global_conf.bptt + 2]:copy_fromd(hidden_store[tnow - global_conf.bptt - 1])
- nerv.LMUtil.set_onehot(dagL_input[global_conf.bptt + 3], token_store[tnow + 1], global_conf.vocab)
+ nerv.LMUtil.set_onehot(dagL_input[global_conf.bptt + 3], token_store[tnow + 1], global_conf.vocab) --for softmax
--local dagL_input = create_dag_input(global_conf, token_store, hidden_store, tnow)
global_conf.timer:tic("dagL-propagate")
@@ -224,6 +224,7 @@ function propagateFile(global_conf, dagL, fn, config)
for key, value in pairs(global_conf.timer.rec) do
printf("\t [global_conf.timer]: time spent on %s:%.5fs\n", key, value)
end
+ --comment this for debughtx
global_conf.timer:flush()
--nerv.CuMatrix.print_profile()
--nerv.CuMatrix.clear_profile()
@@ -277,10 +278,10 @@ if (set == "ptb") then
valid_fn = valid_fn,
test_fn = test_fn,
sche_log_pre = "[SCHEDULER]:",
- log_w_num = 10000, --give a message when log_w_num words have been processed
+ log_w_num = 500000, --give a message when log_w_num words have been processed
timer = nerv.Timer()
}
- global_conf.work_dir = work_dir_base.."/h"..global_conf.hidden_size.."bp"..global_conf.bptt.."slr"..global_conf.lrate..os.date("_%bD%dH%H")
+ global_conf.work_dir = work_dir_base.."/h"..global_conf.hidden_size.."bp"..global_conf.bptt.."slr"..global_conf.lrate --..os.date("_%bD%dH%H") --comment this for testing
global_conf.param_fn = global_conf.work_dir.."/params"
elseif (set == "test") then
train_fn = "/slfs1/users/txh18/workspace/nerv-project/some-text"