aboutsummaryrefslogtreecommitdiff
path: root/nerv/examples/lmptb/m-tests/tnn_test.lua
diff options
context:
space:
mode:
authortxh18 <cloudygooseg@gmail.com>2015-11-12 11:50:07 +0800
committertxh18 <cloudygooseg@gmail.com>2015-11-12 11:50:07 +0800
commitae2bb39ec6ea46a8bdfbd3b8b145ecfb7ca9032f (patch)
tree9a23de71eaf7905a618a6f44cc2316d398376c65 /nerv/examples/lmptb/m-tests/tnn_test.lua
parent73402335834c990dbe6a7729ace7a830ed2f91ae (diff)
get good PPL for h300, see m-tests/LOG-tnn-h300
Diffstat (limited to 'nerv/examples/lmptb/m-tests/tnn_test.lua')
-rw-r--r--nerv/examples/lmptb/m-tests/tnn_test.lua8
1 files changed, 4 insertions, 4 deletions
diff --git a/nerv/examples/lmptb/m-tests/tnn_test.lua b/nerv/examples/lmptb/m-tests/tnn_test.lua
index c4890b6..e2c0d39 100644
--- a/nerv/examples/lmptb/m-tests/tnn_test.lua
+++ b/nerv/examples/lmptb/m-tests/tnn_test.lua
@@ -243,10 +243,10 @@ global_conf = {
mmat_type = nerv.MMatrixFloat,
nn_act_default = 0,
- hidden_size = 200,
- chunk_size = 5,
+ hidden_size = 300,
+ chunk_size = 15,
batch_size = 10,
- max_iter = 20,
+ max_iter = 30,
param_random = function() return (math.random() / 5 - 0.1) end,
train_fn = train_fn,
@@ -322,7 +322,7 @@ for iter = 1, global_conf.max_iter, 1 do
result = lm_process_file(global_conf, global_conf.valid_fn, tnn, false) --false update!
ppl_rec[iter] = result:ppl_net("rnn")
lr_rec[iter] = global_conf.lrate
- if (ppl_last / ppl_rec[iter] < 1.03 or lr_half == true) then
+ if (ppl_last / ppl_rec[iter] < 1.0003 or lr_half == true) then
global_conf.lrate = (global_conf.lrate * 0.6)
lr_half = true
end