From e828c358b2637eac603f519b981737977ef4cb64 Mon Sep 17 00:00:00 2001 From: txh18 Date: Wed, 28 Oct 2015 19:13:30 +0800 Subject: doc change --- README.md | 13 ++++++++++--- 1 file changed, 10 insertions(+), 3 deletions(-) diff --git a/README.md b/README.md index 5926f63..662e710 100644 --- a/README.md +++ b/README.md @@ -3,22 +3,29 @@ NOTE: This readme is obsolete and will be rearranged, for further information, p This user manual will information about how to use __Nerv__ and __Nerv__'s interface. -##How to use## +##How to make and start using## First make sure you have __lua__ and __CUDA__ installed on your computer. __Nerv__ is currently developed via github.You can download and make __Nerv__ by doing the following: ``` cd ~ git clone https://github.com/Nerv-SJTU/nerv.git cd nerv -git clone https://github.com/Nerv-SJTU/nerv-speech.git speech git submodule init && git submodule update make + +#further, if you want the speech modules +git clone https://github.com/Nerv-SJTU/nerv-speech.git speech make speech ``` The `git submodule` command is for the __luajit__ repository inside __Nerv__. Now, you can try to run some example scripts. ``` -./nerv examples/cumatrix_example.lua +./install/bin/nerv examples/cumatrix_example.lua +``` +To get an example of DNN(for ASR) training, run(this requires the speech modules) +You need to be at or (copy files from) `/slfs1`(SJTU speechlab cluster) to get this running. +``` +./install/bin/nerv nerv/examples/asr_trainer.lua nerv/examples/swb_baseline.lua ``` ##How to contribute## -- cgit v1.2.3 From a766983d167c5eb700ff9aaf0ba7e1c4e97a9cf3 Mon Sep 17 00:00:00 2001 From: txh18 Date: Mon, 16 Nov 2015 22:48:08 +0800 Subject: coding style changes --- nerv/layer/affine.lua | 30 +++++++++++++++--------------- nerv/layer/affine_recurrent.lua | 14 +++++++------- 2 files changed, 22 insertions(+), 22 deletions(-) diff --git a/nerv/layer/affine.lua b/nerv/layer/affine.lua index c24af16..c5084c4 100644 --- a/nerv/layer/affine.lua +++ b/nerv/layer/affine.lua @@ -19,29 +19,29 @@ end function MatrixParam:update(gradient) local gconf = self.gconf - if (gconf.momentum > 0) then + if gconf.momentum > 0 then self.correction:add(self.correction, gradient, gconf.momentum, 1.0) -- momentum gain - local mmt_gain = 1.0 / (1.0 - gconf.momentum); + local mmt_gain = 1.0 / (1.0 - gconf.momentum) local n = self.gconf.batch_size * mmt_gain -- perform update - self.trans:add(self.trans, self.correction, 1.0-gconf.lrate*gconf.wcost/gconf.batch_size, -gconf.lrate/n) + self.trans:add(self.trans, self.correction, 1.0 - gconf.lrate * gconf.wcost / gconf.batch_size, - gconf.lrate / n) else - self.trans:add(self.trans, gradient, 1.0-gconf.lrate*gconf.wcost/gconf.batch_size, -gconf.lrate/gconf.batch_size) + self.trans:add(self.trans, gradient, 1.0 - gconf.lrate * gconf.wcost / gconf.batch_size, - gconf.lrate / gconf.batch_size) end end function MatrixParam:updateEI(err, input) local gconf = self.gconf - if (gconf.momentum > 0) then + if gconf.momentum > 0 then self.correction:mul(input, err, 1.0, gconf.momentum, 'T', 'N') -- momentum gain - local mmt_gain = 1.0 / (1.0 - gconf.momentum); + local mmt_gain = 1.0 / (1.0 - gconf.momentum) local n = self.gconf.batch_size * mmt_gain -- perform update - self.trans:add(self.trans, self.correction, 1.0-gconf.lrate*gconf.wcost/gconf.batch_size, -gconf.lrate/n) + self.trans:add(self.trans, self.correction, 1.0 - gconf.lrate * gconf.wcost / gconf.batch_size, - gconf.lrate / n) else - self.trans:mul(input, err, -gconf.lrate/gconf.batch_size, 1.0-gconf.lrate*gconf.wcost/gconf.batch_size, 'T', 'N') + self.trans:mul(input, err, - gconf.lrate / gconf.batch_size, 1.0 - gconf.lrate * gconf.wcost / gconf.batch_size, 'T', 'N') end end @@ -92,20 +92,20 @@ function AffineLayer:batch_resize(batch_size) end function AffineLayer:update(bp_err, input, output) - if (self.direct_update == true) then + if self.direct_update == true then local gconf = self.gconf - if (gconf.momentum > 0) then + if gconf.momentum > 0 then self.ltp.correction:mul(input[1], bp_err[1], 1.0, gconf.momentum, 'T', 'N') self.bp.correction:add(self.bp.correction, bp_err[1]:colsum(), gconf.momentum, 1) -- momentum gain - local mmt_gain = 1.0 / (1.0 - gconf.momentum); + local mmt_gain = 1.0 / (1.0 - gconf.momentum) local n = self.gconf.batch_size * mmt_gain -- perform update - self.ltp.trans:add(self.ltp.trans, self.ltp.correction, 1.0-gconf.lrate*gconf.wcost/gconf.batch_size, -gconf.lrate / n) - self.bp.trans:add(self.bp.trans, self.bp.correction, 1.0-gconf.lrate*gconf.wcost/gconf.batch_size, -gconf.lrate / n) + self.ltp.trans:add(self.ltp.trans, self.ltp.correction, 1.0 - gconf.lrate * gconf.wcost / gconf.batch_size, - gconf.lrate / n) + self.bp.trans:add(self.bp.trans, self.bp.correction, 1.0 - gconf.lrate * gconf.wcost / gconf.batch_size, - gconf.lrate / n) else - self.ltp.trans:mul(input[1], bp_err[1], -gconf.lrate / gconf.batch_size, 1.0-gconf.lrate*gconf.wcost/gconf.batch_size, 'T', 'N') - self.bp.trans:add(self.bp.trans, bp_err[1]:colsum(), 1.0-gconf.lrate*gconf.wcost/gconf.batch_size, -gconf.lrate / gconf.batch_size) + self.ltp.trans:mul(input[1], bp_err[1], - gconf.lrate / gconf.batch_size, 1.0 - gconf.lrate * gconf.wcost / gconf.batch_size, 'T', 'N') + self.bp.trans:add(self.bp.trans, bp_err[1]:colsum(), 1.0 - gconf.lrate * gconf.wcost / gconf.batch_size, - gconf.lrate / gconf.batch_size) end else self.ltp:updateEI(bp_err[1], input[1]) diff --git a/nerv/layer/affine_recurrent.lua b/nerv/layer/affine_recurrent.lua index b465e95..7f9c3f5 100644 --- a/nerv/layer/affine_recurrent.lua +++ b/nerv/layer/affine_recurrent.lua @@ -42,23 +42,23 @@ function Recurrent:batch_resize(batch_size) end function Recurrent:update(bp_err, input, output) - if (self.direct_update == true) then + if self.direct_update == true then local ltp_hh = self.ltp_hh.trans local bp = self.bp.trans local gconf = self.gconf if (gconf.momentum > 0) then -- momentum gain - local mmt_gain = 1.0 / (1.0 - gconf.momentum); + local mmt_gain = 1.0 / (1.0 - gconf.momentum) local n = input[1]:nrow() * mmt_gain -- update corrections (accumulated errors) self.ltp_hh.correction:mul(input[2], bp_err[1], 1.0, gconf.momentum, 'T', 'N') self.bp.correction:add(self.bp.correction, bp_err[1]:colsum(), gconf.momentum, 1.0) -- perform update and weight decay - ltp_hh:add(ltp_hh, self.ltp_hh.correction, 1.0-gconf.lrate*gconf.wcost/gconf.batch_size, -gconf.lrate/n) - bp:add(bp, self.bp.correction, 1.0-gconf.lrate*gconf.wcost/gconf.batch_size, -gconf.lrate/n) + ltp_hh:add(ltp_hh, self.ltp_hh.correction, 1.0 - gconf.lrate * gconf.wcost / gconf.batch_size, - gconf.lrate / n) + bp:add(bp, self.bp.correction, 1.0 - gconf.lrate * gconf.wcost / gconf.batch_size, - gconf.lrate / n) else - ltp_hh:mul(input[2], bp_err[1], -gconf.lrate/gconf.batch_size, 1.0-gconf.wcost*gconf.lrate/gconf.batch_size, 'T', 'N') - bp:add(bp, bp_err[1]:colsum(), 1.0-gconf.lrate*gconf.wcost/gconf.batch_size, -gconf.lrate/gconf.batch_size) + ltp_hh:mul(input[2], bp_err[1], - gconf.lrate / gconf.batch_size, 1.0 - gconf.wcost * gconf.lrate / gconf.batch_size, 'T', 'N') + bp:add(bp, bp_err[1]:colsum(), 1.0 - gconf.lrate * gconf.wcost / gconf.batch_size, - gconf.lrate / gconf.batch_size) end else --self.ltp_hh_grad:mul(input[2], bp_err[1], 1.0, 0.0, 'T', 'N') @@ -85,7 +85,7 @@ function Recurrent:back_propagate(bp_err, next_bp_err, input, output) end ]]-- if (self.clip ~= nil) then - next_bp_err[2]:clip(-self.clip, self.clip) + next_bp_err[2]:clip(- self.clip, self.clip) end end -- cgit v1.2.3 From cbcce5ecc2864872b411eebbd307fa0f9a7e9dc0 Mon Sep 17 00:00:00 2001 From: txh18 Date: Mon, 16 Nov 2015 23:21:43 +0800 Subject: change updateEI to update_by_err_input --- nerv/layer/affine.lua | 8 ++++---- nerv/layer/affine_recurrent.lua | 4 ++-- 2 files changed, 6 insertions(+), 6 deletions(-) diff --git a/nerv/layer/affine.lua b/nerv/layer/affine.lua index c5084c4..3ba9408 100644 --- a/nerv/layer/affine.lua +++ b/nerv/layer/affine.lua @@ -17,7 +17,7 @@ function MatrixParam:train_init() self.correction:fill(0) end -function MatrixParam:update(gradient) +function MatrixParam:update_by_gradient(gradient) local gconf = self.gconf if gconf.momentum > 0 then self.correction:add(self.correction, gradient, gconf.momentum, 1.0) @@ -31,7 +31,7 @@ function MatrixParam:update(gradient) end end -function MatrixParam:updateEI(err, input) +function MatrixParam:update_by_err_input(err, input) local gconf = self.gconf if gconf.momentum > 0 then self.correction:mul(input, err, 1.0, gconf.momentum, 'T', 'N') @@ -108,8 +108,8 @@ function AffineLayer:update(bp_err, input, output) self.bp.trans:add(self.bp.trans, bp_err[1]:colsum(), 1.0 - gconf.lrate * gconf.wcost / gconf.batch_size, - gconf.lrate / gconf.batch_size) end else - self.ltp:updateEI(bp_err[1], input[1]) - self.bp:update(bp_err[1]:colsum()) + self.ltp:update_by_err_input(bp_err[1], input[1]) + self.bp:update_by_gradient(bp_err[1]:colsum()) end end diff --git a/nerv/layer/affine_recurrent.lua b/nerv/layer/affine_recurrent.lua index 7f9c3f5..da189e0 100644 --- a/nerv/layer/affine_recurrent.lua +++ b/nerv/layer/affine_recurrent.lua @@ -62,8 +62,8 @@ function Recurrent:update(bp_err, input, output) end else --self.ltp_hh_grad:mul(input[2], bp_err[1], 1.0, 0.0, 'T', 'N') - self.ltp_hh:updateEI(bp_err[1], input[2]) - self.bp:update(bp_err[1]:colsum()) + self.ltp_hh:update_by_err_input(bp_err[1], input[2]) + self.bp:update_by_gradient(bp_err[1]:colsum()) end end -- cgit v1.2.3 From ad5cb00e4062a84f8c033bc7090a6047669a66c9 Mon Sep 17 00:00:00 2001 From: txh18 Date: Mon, 16 Nov 2015 23:47:20 +0800 Subject: added LOG-tnn-h400 LOG --- nerv/examples/lmptb/m-tests/LOG-tnn-h300 | 3824 ---------------- nerv/examples/lmptb/m-tests/LOG-tnn-h400 | 7193 ++++++++++++++++++++++++++++++ 2 files changed, 7193 insertions(+), 3824 deletions(-) delete mode 100644 nerv/examples/lmptb/m-tests/LOG-tnn-h300 create mode 100644 nerv/examples/lmptb/m-tests/LOG-tnn-h400 diff --git a/nerv/examples/lmptb/m-tests/LOG-tnn-h300 b/nerv/examples/lmptb/m-tests/LOG-tnn-h300 deleted file mode 100644 index ca4ff77..0000000 --- a/nerv/examples/lmptb/m-tests/LOG-tnn-h300 +++ /dev/null @@ -1,3824 +0,0 @@ -Greetings -[SCHEDULER]: printing global_conf -test_fn /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.test.txt.adds -train_fn /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.train.txt.adds -sche_log_pre [SCHEDULER]: -train_fn_shuf_bak /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf_bak -mmat_type table: 0x40f7afc8 -wcost 1e-06 -train_fn_shuf /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf -log_w_num 40000 -param_fn /home/slhome/txh18/workspace/nerv/play/dagL_test/params -batch_size 10 -hidden_size 300 -timer nerv.Timer -work_dir /home/slhome/txh18/workspace/nerv/play/dagL_test -nn_act_default 0 -max_iter 30 -valid_fn /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds -lrate 1 -momentum 0 -param_random function: 0x41ba6b68 -chunk_size 15 -cumat_type table: 0x40f7e4e0 -[SCHEDULER]: creating work_dir... -[LOG]LMVocab: Vocab building on file /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.train.txt.adds... -[LOG]LMVocab: Building finished, vocab size now is 10000. -[SCHEDULER]: preparing parameters... -reading chunk 0 from 0 -metadata: return {type="nerv.LinearTransParam",id="ltp_ih"} - -reading chunk 1 from 34509841 -metadata: return {type="nerv.LinearTransParam",id="ltp_hh"} - -reading chunk 2 from 35545246 -metadata: return {type="nerv.LinearTransParam",id="ltp_ho"} - -reading chunk 3 from 70044639 -metadata: return {type="nerv.BiasParam",id="bp_h"} - -reading chunk 4 from 70048135 -metadata: return {type="nerv.BiasParam",id="bp_o"} - -reading chunk 5 from 70163127 -[SCHEDULER]: preparing parameters end. -===INITIAL VALIDATION=== -[SCHEDULER]: preparing parameters... -reading chunk 0 from 0 -metadata: return {type="nerv.LinearTransParam",id="ltp_ih"} - -reading chunk 1 from 34509841 -metadata: return {type="nerv.LinearTransParam",id="ltp_hh"} - -reading chunk 2 from 35545246 -metadata: return {type="nerv.LinearTransParam",id="ltp_ho"} - -reading chunk 3 from 70044639 -metadata: return {type="nerv.BiasParam",id="bp_h"} - -reading chunk 4 from 70048135 -metadata: return {type="nerv.BiasParam",id="bp_o"} - -reading chunk 5 from 70163127 -[SCHEDULER]: preparing parameters end. -[SCHEDULER]: preparing layers... -(21:26:39 2015-11-12)[nerv] info: create layer: recurrentL1 -(21:26:39 2015-11-12)[nerv] info: create layer: sigmoidL1 -(21:26:39 2015-11-12)[nerv] info: create layer: combinerL1 -(21:26:39 2015-11-12)[nerv] info: create layer: outputL -(21:26:39 2015-11-12)[nerv] info: create layer: softmaxL -(21:26:39 2015-11-12)[nerv] info: create layer: selectL1 -[SCHEDULER]: preparing layers end. -[SCHEDULER]: Generate and initing TNN ... - selectL1 0 -selectL1 recurrentL1 0 -recurrentL1 sigmoidL1 0 -sigmoidL1 combinerL1 0 -combinerL1 recurrentL1 1 -combinerL1 outputL 0 -outputL softmaxL 0 - softmaxL 0 -softmaxL 0 -recurrentL1 #dim_in: 2 #dim_out: 1 #i_conns_p: 2 #o_conns_p 1 -outputL #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -combinerL1 #dim_in: 1 #dim_out: 2 #i_conns_p: 1 #o_conns_p 2 -sigmoidL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -selectL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 0 #o_conns_p 1 -softmaxL #dim_in: 2 #dim_out: 1 #i_conns_p: 1 #o_conns_p 0 -TNN initing storage selectL1 -> recurrentL1 -TNN initing storage recurrentL1 -> sigmoidL1 -TNN initing storage sigmoidL1 -> combinerL1 -TNN initing storage combinerL1 -> recurrentL1 -TNN initing storage combinerL1 -> outputL -TNN initing storage outputL -> softmaxL -[SCHEDULER]: Initing TNN end. -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds... -batch_size: 10 chunk_size 15 -[SCHEDULER]: 40095 words processed Thu Nov 12 21:26:45 2015. - [SCHEDULER]: log prob per sample :-4.052709. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER]: Displaying result: -[SCHEDULER]: LMResult status of rnn: -[SCHEDULER]: Doing on /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds end. - -[SCHEDULER]: preparing parameters... -reading chunk 0 from 0 -metadata: return {type="nerv.LinearTransParam",id="ltp_ih"} - -reading chunk 1 from 34509841 -metadata: return {type="nerv.LinearTransParam",id="ltp_hh"} - -reading chunk 2 from 35545246 -metadata: return {type="nerv.LinearTransParam",id="ltp_ho"} - -reading chunk 3 from 70044639 -metadata: return {type="nerv.BiasParam",id="bp_h"} - -reading chunk 4 from 70048135 -metadata: return {type="nerv.BiasParam",id="bp_o"} - -reading chunk 5 from 70163127 -[SCHEDULER]: preparing parameters end. -[SCHEDULER]: preparing layers... -(21:26:55 2015-11-12)[nerv] info: create layer: recurrentL1 -(21:26:55 2015-11-12)[nerv] info: create layer: sigmoidL1 -(21:26:55 2015-11-12)[nerv] info: create layer: combinerL1 -(21:26:55 2015-11-12)[nerv] info: create layer: outputL -(21:26:55 2015-11-12)[nerv] info: create layer: softmaxL -(21:26:55 2015-11-12)[nerv] info: create layer: selectL1 -[SCHEDULER]: preparing layers end. -[SCHEDULER]: Generate and initing TNN ... - selectL1 0 -selectL1 recurrentL1 0 -recurrentL1 sigmoidL1 0 -sigmoidL1 combinerL1 0 -combinerL1 recurrentL1 1 -combinerL1 outputL 0 -outputL softmaxL 0 - softmaxL 0 -softmaxL 0 -recurrentL1 #dim_in: 2 #dim_out: 1 #i_conns_p: 2 #o_conns_p 1 -outputL #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -combinerL1 #dim_in: 1 #dim_out: 2 #i_conns_p: 1 #o_conns_p 2 -sigmoidL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -selectL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 0 #o_conns_p 1 -softmaxL #dim_in: 2 #dim_out: 1 #i_conns_p: 1 #o_conns_p 0 -TNN initing storage selectL1 -> recurrentL1 -TNN initing storage recurrentL1 -> sigmoidL1 -TNN initing storage sigmoidL1 -> combinerL1 -TNN initing storage combinerL1 -> recurrentL1 -TNN initing storage combinerL1 -> outputL -TNN initing storage outputL -> softmaxL -[SCHEDULER]: Initing TNN end. -===ITERATION 1 LR 1.000000=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER1 LR1]: 40099 words processed Thu Nov 12 21:27:11 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-3.292692. -[SCHEDULER ITER1 LR1]: 80063 words processed Thu Nov 12 21:27:27 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-3.070144. -[SCHEDULER ITER1 LR1]: 120068 words processed Thu Nov 12 21:27:43 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.984966. -[SCHEDULER ITER1 LR1]: 160017 words processed Thu Nov 12 21:27:59 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.924306. -[SCHEDULER ITER1 LR1]: 200138 words processed Thu Nov 12 21:28:15 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.895537. -[SCHEDULER ITER1 LR1]: 240007 words processed Thu Nov 12 21:28:31 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.866355. -[SCHEDULER ITER1 LR1]: 280135 words processed Thu Nov 12 21:28:47 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.841706. -[SCHEDULER ITER1 LR1]: 320080 words processed Thu Nov 12 21:29:03 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.820864. -[SCHEDULER ITER1 LR1]: 360059 words processed Thu Nov 12 21:29:19 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.798327. -[SCHEDULER ITER1 LR1]: 400021 words processed Thu Nov 12 21:29:36 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.784042. -[SCHEDULER ITER1 LR1]: 440102 words processed Thu Nov 12 21:29:53 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.767230. -[SCHEDULER ITER1 LR1]: 480051 words processed Thu Nov 12 21:30:09 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.750979. -[SCHEDULER ITER1 LR1]: 520093 words processed Thu Nov 12 21:30:25 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.737316. -[SCHEDULER ITER1 LR1]: 560039 words processed Thu Nov 12 21:30:41 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.721875. -[SCHEDULER ITER1 LR1]: 600112 words processed Thu Nov 12 21:30:57 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.712290. -[SCHEDULER ITER1 LR1]: 640076 words processed Thu Nov 12 21:31:13 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.701624. -[SCHEDULER ITER1 LR1]: 680026 words processed Thu Nov 12 21:31:29 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.692884. -[SCHEDULER ITER1 LR1]: 720133 words processed Thu Nov 12 21:31:45 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.684241. -[SCHEDULER ITER1 LR1]: 760048 words processed Thu Nov 12 21:32:01 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.673189. -[SCHEDULER ITER1 LR1]: 800117 words processed Thu Nov 12 21:32:17 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.665760. -[SCHEDULER ITER1 LR1]: 840116 words processed Thu Nov 12 21:32:33 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.656155. -[SCHEDULER ITER1 LR1]: 880037 words processed Thu Nov 12 21:32:49 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.648768. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER1 LR1]: Displaying result: -[SCHEDULER ITER1 LR1]: LMResult status of rnn: -[SCHEDULER ITER1 LR1]: Doing on /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf end. -[SCHEDULER ITER1 LR1]: shuffling training file -===PEEK ON TEST 1=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.test.txt.adds... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER1 LR1]: 40087 words processed Thu Nov 12 21:33:02 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.504355. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER1 LR1]: Displaying result: -[SCHEDULER ITER1 LR1]: LMResult status of rnn: -[SCHEDULER ITER1 LR1]: Doing on /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.test.txt.adds end. -===VALIDATION 1=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER1 LR1]: 40095 words processed Thu Nov 12 21:33:14 2015. - [SCHEDULER ITER1 LR1]: log prob per sample :-2.550268. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER1 LR1]: Displaying result: -[SCHEDULER ITER1 LR1]: LMResult status of rnn: -[SCHEDULER ITER1 LR1]: Doing on /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds end. -[SCHEDULER ITER1 LR1]: saving net to file /home/slhome/txh18/workspace/nerv/play/dagL_test/params... - -[SCHEDULER ITER1 LR1]: preparing parameters... -reading chunk 0 from 0 -metadata: return {type="nerv.LinearTransParam",id="ltp_hh"} - -reading chunk 1 from 1075976 -metadata: return {type="nerv.BiasParam",id="bp_h"} - -reading chunk 2 from 1079632 -metadata: return {type="nerv.LinearTransParam",id="ltp_ho"} - -reading chunk 3 from 35607920 -metadata: return {type="nerv.BiasParam",id="bp_o"} - -reading chunk 4 from 35724250 -metadata: return {type="nerv.LinearTransParam",id="ltp_ih"} - -reading chunk 5 from 70241455 -[SCHEDULER ITER1 LR1]: preparing parameters end. -[SCHEDULER ITER1 LR1]: preparing layers... -(21:33:24 2015-11-12)[nerv] info: create layer: recurrentL1 -(21:33:24 2015-11-12)[nerv] info: create layer: sigmoidL1 -(21:33:24 2015-11-12)[nerv] info: create layer: combinerL1 -(21:33:24 2015-11-12)[nerv] info: create layer: outputL -(21:33:24 2015-11-12)[nerv] info: create layer: softmaxL -(21:33:24 2015-11-12)[nerv] info: create layer: selectL1 -[SCHEDULER ITER1 LR1]: preparing layers end. -[SCHEDULER ITER1 LR1]: Generate and initing TNN ... - selectL1 0 -selectL1 recurrentL1 0 -recurrentL1 sigmoidL1 0 -sigmoidL1 combinerL1 0 -combinerL1 recurrentL1 1 -combinerL1 outputL 0 -outputL softmaxL 0 - softmaxL 0 -softmaxL 0 -recurrentL1 #dim_in: 2 #dim_out: 1 #i_conns_p: 2 #o_conns_p 1 -outputL #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -combinerL1 #dim_in: 1 #dim_out: 2 #i_conns_p: 1 #o_conns_p 2 -sigmoidL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -selectL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 0 #o_conns_p 1 -softmaxL #dim_in: 2 #dim_out: 1 #i_conns_p: 1 #o_conns_p 0 -TNN initing storage selectL1 -> recurrentL1 -TNN initing storage recurrentL1 -> sigmoidL1 -TNN initing storage sigmoidL1 -> combinerL1 -TNN initing storage combinerL1 -> recurrentL1 -TNN initing storage combinerL1 -> outputL -TNN initing storage outputL -> softmaxL -[SCHEDULER ITER1 LR1]: Initing TNN end. -===ITERATION 2 LR 1.000000=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER2 LR1]: 40092 words processed Thu Nov 12 21:33:40 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.611210. -[SCHEDULER ITER2 LR1]: 80099 words processed Thu Nov 12 21:33:56 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.589139. -[SCHEDULER ITER2 LR1]: 120004 words processed Thu Nov 12 21:34:12 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.570733. -[SCHEDULER ITER2 LR1]: 160114 words processed Thu Nov 12 21:34:28 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.551667. -[SCHEDULER ITER2 LR1]: 200066 words processed Thu Nov 12 21:34:44 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.540568. -[SCHEDULER ITER2 LR1]: 240045 words processed Thu Nov 12 21:35:00 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.526721. -[SCHEDULER ITER2 LR1]: 280057 words processed Thu Nov 12 21:35:16 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.517728. -[SCHEDULER ITER2 LR1]: 320106 words processed Thu Nov 12 21:35:32 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.508229. -[SCHEDULER ITER2 LR1]: 360024 words processed Thu Nov 12 21:35:48 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.502058. -[SCHEDULER ITER2 LR1]: 400089 words processed Thu Nov 12 21:36:04 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.497895. -[SCHEDULER ITER2 LR1]: 440067 words processed Thu Nov 12 21:36:20 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.493752. -[SCHEDULER ITER2 LR1]: 480051 words processed Thu Nov 12 21:36:36 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.490189. -[SCHEDULER ITER2 LR1]: 520140 words processed Thu Nov 12 21:36:52 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.486542. -[SCHEDULER ITER2 LR1]: 560132 words processed Thu Nov 12 21:37:08 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.483445. -[SCHEDULER ITER2 LR1]: 600118 words processed Thu Nov 12 21:37:24 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.483056. -[SCHEDULER ITER2 LR1]: 640090 words processed Thu Nov 12 21:37:40 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.480205. -[SCHEDULER ITER2 LR1]: 680075 words processed Thu Nov 12 21:37:56 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.478075. -[SCHEDULER ITER2 LR1]: 720043 words processed Thu Nov 12 21:38:12 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.475920. -[SCHEDULER ITER2 LR1]: 760012 words processed Thu Nov 12 21:38:28 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.473641. -[SCHEDULER ITER2 LR1]: 800113 words processed Thu Nov 12 21:38:44 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.471283. -[SCHEDULER ITER2 LR1]: 840089 words processed Thu Nov 12 21:39:00 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.468366. -[SCHEDULER ITER2 LR1]: 880052 words processed Thu Nov 12 21:39:16 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.466238. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER2 LR1]: Displaying result: -[SCHEDULER ITER2 LR1]: LMResult status of rnn: -[SCHEDULER ITER2 LR1]: Doing on /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf end. -[SCHEDULER ITER2 LR1]: shuffling training file -===PEEK ON TEST 2=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.test.txt.adds... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER2 LR1]: 40087 words processed Thu Nov 12 21:39:27 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.378341. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER2 LR1]: Displaying result: -[SCHEDULER ITER2 LR1]: LMResult status of rnn: -[SCHEDULER ITER2 LR1]: Doing on /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.test.txt.adds end. -===VALIDATION 2=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER2 LR1]: 40095 words processed Thu Nov 12 21:39:39 2015. - [SCHEDULER ITER2 LR1]: log prob per sample :-2.425320. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER2 LR1]: Displaying result: -[SCHEDULER ITER2 LR1]: LMResult status of rnn: -[SCHEDULER ITER2 LR1]: Doing on /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds end. -[SCHEDULER ITER2 LR1]: saving net to file /home/slhome/txh18/workspace/nerv/play/dagL_test/params... - -[SCHEDULER ITER2 LR1]: preparing parameters... -reading chunk 0 from 0 -metadata: return {type="nerv.LinearTransParam",id="ltp_ih"} - -reading chunk 1 from 34522576 -metadata: return {type="nerv.BiasParam",id="bp_h"} - -reading chunk 2 from 34526234 -metadata: return {type="nerv.LinearTransParam",id="ltp_ho"} - -reading chunk 3 from 69072330 -metadata: return {type="nerv.BiasParam",id="bp_o"} - -reading chunk 4 from 69188461 -metadata: return {type="nerv.LinearTransParam",id="ltp_hh"} - -reading chunk 5 from 70263917 -[SCHEDULER ITER2 LR1]: preparing parameters end. -[SCHEDULER ITER2 LR1]: preparing layers... -(21:39:50 2015-11-12)[nerv] info: create layer: recurrentL1 -(21:39:50 2015-11-12)[nerv] info: create layer: sigmoidL1 -(21:39:50 2015-11-12)[nerv] info: create layer: combinerL1 -(21:39:50 2015-11-12)[nerv] info: create layer: outputL -(21:39:50 2015-11-12)[nerv] info: create layer: softmaxL -(21:39:50 2015-11-12)[nerv] info: create layer: selectL1 -[SCHEDULER ITER2 LR1]: preparing layers end. -[SCHEDULER ITER2 LR1]: Generate and initing TNN ... - selectL1 0 -selectL1 recurrentL1 0 -recurrentL1 sigmoidL1 0 -sigmoidL1 combinerL1 0 -combinerL1 recurrentL1 1 -combinerL1 outputL 0 -outputL softmaxL 0 - softmaxL 0 -softmaxL 0 -recurrentL1 #dim_in: 2 #dim_out: 1 #i_conns_p: 2 #o_conns_p 1 -outputL #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -combinerL1 #dim_in: 1 #dim_out: 2 #i_conns_p: 1 #o_conns_p 2 -sigmoidL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -selectL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 0 #o_conns_p 1 -softmaxL #dim_in: 2 #dim_out: 1 #i_conns_p: 1 #o_conns_p 0 -TNN initing storage selectL1 -> recurrentL1 -TNN initing storage recurrentL1 -> sigmoidL1 -TNN initing storage sigmoidL1 -> combinerL1 -TNN initing storage combinerL1 -> recurrentL1 -TNN initing storage combinerL1 -> outputL -TNN initing storage outputL -> softmaxL -[SCHEDULER ITER2 LR1]: Initing TNN end. -===ITERATION 3 LR 1.000000=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER3 LR1]: 40092 words processed Thu Nov 12 21:40:06 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.544465. -[SCHEDULER ITER3 LR1]: 80099 words processed Thu Nov 12 21:40:22 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.523845. -[SCHEDULER ITER3 LR1]: 120004 words processed Thu Nov 12 21:40:38 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.503563. -[SCHEDULER ITER3 LR1]: 160114 words processed Thu Nov 12 21:40:55 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.492793. -[SCHEDULER ITER3 LR1]: 200066 words processed Thu Nov 12 21:41:11 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.486434. -[SCHEDULER ITER3 LR1]: 240045 words processed Thu Nov 12 21:41:27 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.477573. -[SCHEDULER ITER3 LR1]: 280057 words processed Thu Nov 12 21:41:43 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.472171. -[SCHEDULER ITER3 LR1]: 320106 words processed Thu Nov 12 21:41:59 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.464689. -[SCHEDULER ITER3 LR1]: 360024 words processed Thu Nov 12 21:42:15 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.462254. -[SCHEDULER ITER3 LR1]: 400089 words processed Thu Nov 12 21:42:31 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.460639. -[SCHEDULER ITER3 LR1]: 440067 words processed Thu Nov 12 21:42:47 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.457377. -[SCHEDULER ITER3 LR1]: 480051 words processed Thu Nov 12 21:43:03 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.454522. -[SCHEDULER ITER3 LR1]: 520140 words processed Thu Nov 12 21:43:20 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.451702. -[SCHEDULER ITER3 LR1]: 560132 words processed Thu Nov 12 21:43:36 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.449800. -[SCHEDULER ITER3 LR1]: 600118 words processed Thu Nov 12 21:43:52 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.460651. -[SCHEDULER ITER3 LR1]: 640090 words processed Thu Nov 12 21:44:08 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.478514. -[SCHEDULER ITER3 LR1]: 680075 words processed Thu Nov 12 21:44:24 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.481904. -[SCHEDULER ITER3 LR1]: 720043 words processed Thu Nov 12 21:44:40 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.481315. -[SCHEDULER ITER3 LR1]: 760012 words processed Thu Nov 12 21:44:56 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.479536. -[SCHEDULER ITER3 LR1]: 800113 words processed Thu Nov 12 21:45:12 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.477516. -[SCHEDULER ITER3 LR1]: 840089 words processed Thu Nov 12 21:45:28 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.474938. -[SCHEDULER ITER3 LR1]: 880052 words processed Thu Nov 12 21:45:44 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.472685. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER3 LR1]: Displaying result: -[SCHEDULER ITER3 LR1]: LMResult status of rnn: -[SCHEDULER ITER3 LR1]: Doing on /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf end. -[SCHEDULER ITER3 LR1]: shuffling training file -===PEEK ON TEST 3=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.test.txt.adds... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER3 LR1]: 40087 words processed Thu Nov 12 21:45:55 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.372994. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER3 LR1]: Displaying result: -[SCHEDULER ITER3 LR1]: LMResult status of rnn: -[SCHEDULER ITER3 LR1]: Doing on /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.test.txt.adds end. -===VALIDATION 3=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER3 LR1]: 40095 words processed Thu Nov 12 21:46:07 2015. - [SCHEDULER ITER3 LR1]: log prob per sample :-2.423095. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER3 LR1]: Displaying result: -[SCHEDULER ITER3 LR1]: LMResult status of rnn: -[SCHEDULER ITER3 LR1]: Doing on /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds end. -[SCHEDULER ITER3 LR1]: saving net to file /home/slhome/txh18/workspace/nerv/play/dagL_test/params... - -[SCHEDULER ITER3 LR1]: preparing parameters... -reading chunk 0 from 0 -metadata: return {type="nerv.LinearTransParam",id="ltp_hh"} - -reading chunk 1 from 1075055 -metadata: return {type="nerv.BiasParam",id="bp_h"} - -reading chunk 2 from 1078714 -metadata: return {type="nerv.LinearTransParam",id="ltp_ho"} - -reading chunk 3 from 35640630 -metadata: return {type="nerv.BiasParam",id="bp_o"} - -reading chunk 4 from 35756784 -metadata: return {type="nerv.LinearTransParam",id="ltp_ih"} - -reading chunk 5 from 70282927 -[SCHEDULER ITER3 LR1]: preparing parameters end. -[SCHEDULER ITER3 LR1]: preparing layers... -(21:46:18 2015-11-12)[nerv] info: create layer: recurrentL1 -(21:46:18 2015-11-12)[nerv] info: create layer: sigmoidL1 -(21:46:18 2015-11-12)[nerv] info: create layer: combinerL1 -(21:46:18 2015-11-12)[nerv] info: create layer: outputL -(21:46:18 2015-11-12)[nerv] info: create layer: softmaxL -(21:46:18 2015-11-12)[nerv] info: create layer: selectL1 -[SCHEDULER ITER3 LR1]: preparing layers end. -[SCHEDULER ITER3 LR1]: Generate and initing TNN ... - selectL1 0 -selectL1 recurrentL1 0 -recurrentL1 sigmoidL1 0 -sigmoidL1 combinerL1 0 -combinerL1 recurrentL1 1 -combinerL1 outputL 0 -outputL softmaxL 0 - softmaxL 0 -softmaxL 0 -recurrentL1 #dim_in: 2 #dim_out: 1 #i_conns_p: 2 #o_conns_p 1 -outputL #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -combinerL1 #dim_in: 1 #dim_out: 2 #i_conns_p: 1 #o_conns_p 2 -sigmoidL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -selectL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 0 #o_conns_p 1 -softmaxL #dim_in: 2 #dim_out: 1 #i_conns_p: 1 #o_conns_p 0 -TNN initing storage selectL1 -> recurrentL1 -TNN initing storage recurrentL1 -> sigmoidL1 -TNN initing storage sigmoidL1 -> combinerL1 -TNN initing storage combinerL1 -> recurrentL1 -TNN initing storage combinerL1 -> outputL -TNN initing storage outputL -> softmaxL -[SCHEDULER ITER3 LR1]: Initing TNN end. -===ITERATION 4 LR 1.000000=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER4 LR1]: 40092 words processed Thu Nov 12 21:46:34 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.405192. -[SCHEDULER ITER4 LR1]: 80099 words processed Thu Nov 12 21:46:50 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.387086. -[SCHEDULER ITER4 LR1]: 120004 words processed Thu Nov 12 21:47:06 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.375429. -[SCHEDULER ITER4 LR1]: 160114 words processed Thu Nov 12 21:47:22 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.371996. -[SCHEDULER ITER4 LR1]: 200066 words processed Thu Nov 12 21:47:38 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.369614. -[SCHEDULER ITER4 LR1]: 240045 words processed Thu Nov 12 21:47:54 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.363105. -[SCHEDULER ITER4 LR1]: 280057 words processed Thu Nov 12 21:48:10 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.358276. -[SCHEDULER ITER4 LR1]: 320106 words processed Thu Nov 12 21:48:26 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.352905. -[SCHEDULER ITER4 LR1]: 360024 words processed Thu Nov 12 21:48:42 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.350189. -[SCHEDULER ITER4 LR1]: 400089 words processed Thu Nov 12 21:48:58 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.348116. -[SCHEDULER ITER4 LR1]: 440067 words processed Thu Nov 12 21:49:14 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.346662. -[SCHEDULER ITER4 LR1]: 480051 words processed Thu Nov 12 21:49:30 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.345278. -[SCHEDULER ITER4 LR1]: 520140 words processed Thu Nov 12 21:49:46 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.343802. -[SCHEDULER ITER4 LR1]: 560132 words processed Thu Nov 12 21:50:02 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.342152. -[SCHEDULER ITER4 LR1]: 600118 words processed Thu Nov 12 21:50:18 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.344444. -[SCHEDULER ITER4 LR1]: 640090 words processed Thu Nov 12 21:50:34 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.342672. -[SCHEDULER ITER4 LR1]: 680075 words processed Thu Nov 12 21:50:50 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.341851. -[SCHEDULER ITER4 LR1]: 720043 words processed Thu Nov 12 21:51:06 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.340724. -[SCHEDULER ITER4 LR1]: 760012 words processed Thu Nov 12 21:51:22 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.339464. -[SCHEDULER ITER4 LR1]: 800113 words processed Thu Nov 12 21:51:38 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.338453. -[SCHEDULER ITER4 LR1]: 840089 words processed Thu Nov 12 21:51:53 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.336619. -[SCHEDULER ITER4 LR1]: 880052 words processed Thu Nov 12 21:52:09 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.335602. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER4 LR1]: Displaying result: -[SCHEDULER ITER4 LR1]: LMResult status of rnn: -[SCHEDULER ITER4 LR1]: Doing on /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf end. -[SCHEDULER ITER4 LR1]: shuffling training file -===PEEK ON TEST 4=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.test.txt.adds... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER4 LR1]: 40087 words processed Thu Nov 12 21:52:20 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.303876. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER4 LR1]: Displaying result: -[SCHEDULER ITER4 LR1]: LMResult status of rnn: -[SCHEDULER ITER4 LR1]: Doing on /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.test.txt.adds end. -===VALIDATION 4=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER4 LR1]: 40095 words processed Thu Nov 12 21:52:31 2015. - [SCHEDULER ITER4 LR1]: log prob per sample :-2.356798. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER4 LR1]: Displaying result: -[SCHEDULER ITER4 LR1]: LMResult status of rnn: -[SCHEDULER ITER4 LR1]: Doing on /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds end. -[SCHEDULER ITER4 LR1]: saving net to file /home/slhome/txh18/workspace/nerv/play/dagL_test/params... - -[SCHEDULER ITER4 LR1]: preparing parameters... -reading chunk 0 from 0 -metadata: return {type="nerv.LinearTransParam",id="ltp_ih"} - -reading chunk 1 from 34528542 -metadata: return {type="nerv.BiasParam",id="bp_h"} - -reading chunk 2 from 34532200 -metadata: return {type="nerv.LinearTransParam",id="ltp_ho"} - -reading chunk 3 from 69109612 -metadata: return {type="nerv.BiasParam",id="bp_o"} - -reading chunk 4 from 69225892 -metadata: return {type="nerv.LinearTransParam",id="ltp_hh"} - -reading chunk 5 from 70300614 -[SCHEDULER ITER4 LR1]: preparing parameters end. -[SCHEDULER ITER4 LR1]: preparing layers... -(21:52:41 2015-11-12)[nerv] info: create layer: recurrentL1 -(21:52:41 2015-11-12)[nerv] info: create layer: sigmoidL1 -(21:52:41 2015-11-12)[nerv] info: create layer: combinerL1 -(21:52:41 2015-11-12)[nerv] info: create layer: outputL -(21:52:41 2015-11-12)[nerv] info: create layer: softmaxL -(21:52:41 2015-11-12)[nerv] info: create layer: selectL1 -[SCHEDULER ITER4 LR1]: preparing layers end. -[SCHEDULER ITER4 LR1]: Generate and initing TNN ... - selectL1 0 -selectL1 recurrentL1 0 -recurrentL1 sigmoidL1 0 -sigmoidL1 combinerL1 0 -combinerL1 recurrentL1 1 -combinerL1 outputL 0 -outputL softmaxL 0 - softmaxL 0 -softmaxL 0 -recurrentL1 #dim_in: 2 #dim_out: 1 #i_conns_p: 2 #o_conns_p 1 -outputL #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -combinerL1 #dim_in: 1 #dim_out: 2 #i_conns_p: 1 #o_conns_p 2 -sigmoidL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -selectL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 0 #o_conns_p 1 -softmaxL #dim_in: 2 #dim_out: 1 #i_conns_p: 1 #o_conns_p 0 -TNN initing storage selectL1 -> recurrentL1 -TNN initing storage recurrentL1 -> sigmoidL1 -TNN initing storage sigmoidL1 -> combinerL1 -TNN initing storage combinerL1 -> recurrentL1 -TNN initing storage combinerL1 -> outputL -TNN initing storage outputL -> softmaxL -[SCHEDULER ITER4 LR1]: Initing TNN end. -===ITERATION 5 LR 1.000000=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER5 LR1]: 40092 words processed Thu Nov 12 21:52:56 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.446345. -[SCHEDULER ITER5 LR1]: 80099 words processed Thu Nov 12 21:53:11 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.403749. -[SCHEDULER ITER5 LR1]: 120004 words processed Thu Nov 12 21:53:27 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.384431. -[SCHEDULER ITER5 LR1]: 160114 words processed Thu Nov 12 21:53:43 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.374759. -[SCHEDULER ITER5 LR1]: 200066 words processed Thu Nov 12 21:53:58 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.361125. -[SCHEDULER ITER5 LR1]: 240045 words processed Thu Nov 12 21:54:14 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.347698. -[SCHEDULER ITER5 LR1]: 280057 words processed Thu Nov 12 21:54:30 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.337251. -[SCHEDULER ITER5 LR1]: 320106 words processed Thu Nov 12 21:54:46 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.328378. -[SCHEDULER ITER5 LR1]: 360024 words processed Thu Nov 12 21:55:02 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.322798. -[SCHEDULER ITER5 LR1]: 400089 words processed Thu Nov 12 21:55:18 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.318066. -[SCHEDULER ITER5 LR1]: 440067 words processed Thu Nov 12 21:55:34 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.314987. -[SCHEDULER ITER5 LR1]: 480051 words processed Thu Nov 12 21:55:50 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.312277. -[SCHEDULER ITER5 LR1]: 520140 words processed Thu Nov 12 21:56:06 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.309665. -[SCHEDULER ITER5 LR1]: 560132 words processed Thu Nov 12 21:56:22 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.307166. -[SCHEDULER ITER5 LR1]: 600118 words processed Thu Nov 12 21:56:38 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.304912. -[SCHEDULER ITER5 LR1]: 640090 words processed Thu Nov 12 21:56:54 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.302402. -[SCHEDULER ITER5 LR1]: 680075 words processed Thu Nov 12 21:57:10 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.301702. -[SCHEDULER ITER5 LR1]: 720043 words processed Thu Nov 12 21:57:26 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.300632. -[SCHEDULER ITER5 LR1]: 760012 words processed Thu Nov 12 21:57:42 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.299094. -[SCHEDULER ITER5 LR1]: 800113 words processed Thu Nov 12 21:57:58 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.298026. -[SCHEDULER ITER5 LR1]: 840089 words processed Thu Nov 12 21:58:14 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.295902. -[SCHEDULER ITER5 LR1]: 880052 words processed Thu Nov 12 21:58:30 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.294821. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER5 LR1]: Displaying result: -[SCHEDULER ITER5 LR1]: LMResult status of rnn: -[SCHEDULER ITER5 LR1]: Doing on /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf end. -[SCHEDULER ITER5 LR1]: shuffling training file -===PEEK ON TEST 5=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.test.txt.adds... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER5 LR1]: 40087 words processed Thu Nov 12 21:58:41 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.273549. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER5 LR1]: Displaying result: -[SCHEDULER ITER5 LR1]: LMResult status of rnn: -[SCHEDULER ITER5 LR1]: Doing on /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.test.txt.adds end. -===VALIDATION 5=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER5 LR1]: 40095 words processed Thu Nov 12 21:58:53 2015. - [SCHEDULER ITER5 LR1]: log prob per sample :-2.329139. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER5 LR1]: Displaying result: -[SCHEDULER ITER5 LR1]: LMResult status of rnn: -[SCHEDULER ITER5 LR1]: Doing on /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds end. -[SCHEDULER ITER5 LR1]: saving net to file /home/slhome/txh18/workspace/nerv/play/dagL_test/params... - -[SCHEDULER ITER5 LR1]: preparing parameters... -reading chunk 0 from 0 -metadata: return {type="nerv.LinearTransParam",id="ltp_hh"} - -reading chunk 1 from 1074285 -metadata: return {type="nerv.BiasParam",id="bp_h"} - -reading chunk 2 from 1077942 -metadata: return {type="nerv.LinearTransParam",id="ltp_ho"} - -reading chunk 3 from 35667973 -metadata: return {type="nerv.BiasParam",id="bp_o"} - -reading chunk 4 from 35784319 -metadata: return {type="nerv.LinearTransParam",id="ltp_ih"} - -reading chunk 5 from 70315647 -[SCHEDULER ITER5 LR1]: preparing parameters end. -[SCHEDULER ITER5 LR1]: preparing layers... -(21:59:03 2015-11-12)[nerv] info: create layer: recurrentL1 -(21:59:03 2015-11-12)[nerv] info: create layer: sigmoidL1 -(21:59:03 2015-11-12)[nerv] info: create layer: combinerL1 -(21:59:03 2015-11-12)[nerv] info: create layer: outputL -(21:59:03 2015-11-12)[nerv] info: create layer: softmaxL -(21:59:03 2015-11-12)[nerv] info: create layer: selectL1 -[SCHEDULER ITER5 LR1]: preparing layers end. -[SCHEDULER ITER5 LR1]: Generate and initing TNN ... - selectL1 0 -selectL1 recurrentL1 0 -recurrentL1 sigmoidL1 0 -sigmoidL1 combinerL1 0 -combinerL1 recurrentL1 1 -combinerL1 outputL 0 -outputL softmaxL 0 - softmaxL 0 -softmaxL 0 -recurrentL1 #dim_in: 2 #dim_out: 1 #i_conns_p: 2 #o_conns_p 1 -outputL #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -combinerL1 #dim_in: 1 #dim_out: 2 #i_conns_p: 1 #o_conns_p 2 -sigmoidL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -selectL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 0 #o_conns_p 1 -softmaxL #dim_in: 2 #dim_out: 1 #i_conns_p: 1 #o_conns_p 0 -TNN initing storage selectL1 -> recurrentL1 -TNN initing storage recurrentL1 -> sigmoidL1 -TNN initing storage sigmoidL1 -> combinerL1 -TNN initing storage combinerL1 -> recurrentL1 -TNN initing storage combinerL1 -> outputL -TNN initing storage outputL -> softmaxL -[SCHEDULER ITER5 LR1]: Initing TNN end. -===ITERATION 6 LR 1.000000=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER6 LR1]: 40092 words processed Thu Nov 12 21:59:18 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.302353. -[SCHEDULER ITER6 LR1]: 80099 words processed Thu Nov 12 21:59:34 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.286055. -[SCHEDULER ITER6 LR1]: 120004 words processed Thu Nov 12 21:59:50 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.274566. -[SCHEDULER ITER6 LR1]: 160114 words processed Thu Nov 12 22:00:06 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.272892. -[SCHEDULER ITER6 LR1]: 200066 words processed Thu Nov 12 22:00:22 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.271265. -[SCHEDULER ITER6 LR1]: 240045 words processed Thu Nov 12 22:00:38 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.267254. -[SCHEDULER ITER6 LR1]: 280057 words processed Thu Nov 12 22:00:54 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.263588. -[SCHEDULER ITER6 LR1]: 320106 words processed Thu Nov 12 22:01:10 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.259553. -[SCHEDULER ITER6 LR1]: 360024 words processed Thu Nov 12 22:01:26 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.257404. -[SCHEDULER ITER6 LR1]: 400089 words processed Thu Nov 12 22:01:42 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.255302. -[SCHEDULER ITER6 LR1]: 440067 words processed Thu Nov 12 22:01:58 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.254638. -[SCHEDULER ITER6 LR1]: 480051 words processed Thu Nov 12 22:02:14 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.254091. -[SCHEDULER ITER6 LR1]: 520140 words processed Thu Nov 12 22:02:30 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.253283. -[SCHEDULER ITER6 LR1]: 560132 words processed Thu Nov 12 22:02:46 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.252240. -[SCHEDULER ITER6 LR1]: 600118 words processed Thu Nov 12 22:03:02 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.249917. -[SCHEDULER ITER6 LR1]: 640090 words processed Thu Nov 12 22:03:18 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.250291. -[SCHEDULER ITER6 LR1]: 680075 words processed Thu Nov 12 22:03:34 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.249598. -[SCHEDULER ITER6 LR1]: 720043 words processed Thu Nov 12 22:03:50 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.249209. -[SCHEDULER ITER6 LR1]: 760012 words processed Thu Nov 12 22:04:06 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.248536. -[SCHEDULER ITER6 LR1]: 800113 words processed Thu Nov 12 22:04:22 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.248296. -[SCHEDULER ITER6 LR1]: 840089 words processed Thu Nov 12 22:04:38 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.247162. -[SCHEDULER ITER6 LR1]: 880052 words processed Thu Nov 12 22:04:54 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.246722. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER6 LR1]: Displaying result: -[SCHEDULER ITER6 LR1]: LMResult status of rnn: -[SCHEDULER ITER6 LR1]: Doing on /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf end. -[SCHEDULER ITER6 LR1]: shuffling training file -===PEEK ON TEST 6=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.test.txt.adds... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER6 LR1]: 40087 words processed Thu Nov 12 22:05:06 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.252489. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER6 LR1]: Displaying result: -[SCHEDULER ITER6 LR1]: LMResult status of rnn: -[SCHEDULER ITER6 LR1]: Doing on /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.test.txt.adds end. -===VALIDATION 6=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER6 LR1]: 40095 words processed Thu Nov 12 22:05:18 2015. - [SCHEDULER ITER6 LR1]: log prob per sample :-2.311040. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER6 LR1]: Displaying result: -[SCHEDULER ITER6 LR1]: LMResult status of rnn: -[SCHEDULER ITER6 LR1]: Doing on /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds end. -[SCHEDULER ITER6 LR1]: saving net to file /home/slhome/txh18/workspace/nerv/play/dagL_test/params... - -[SCHEDULER ITER6 LR1]: preparing parameters... -reading chunk 0 from 0 -metadata: return {type="nerv.LinearTransParam",id="ltp_ih"} - -reading chunk 1 from 34533450 -metadata: return {type="nerv.BiasParam",id="bp_h"} - -reading chunk 2 from 34537108 -metadata: return {type="nerv.LinearTransParam",id="ltp_ho"} - -reading chunk 3 from 69139044 -metadata: return {type="nerv.BiasParam",id="bp_o"} - -reading chunk 4 from 69255402 -metadata: return {type="nerv.LinearTransParam",id="ltp_hh"} - -reading chunk 5 from 70329197 -[SCHEDULER ITER6 LR1]: preparing parameters end. -[SCHEDULER ITER6 LR1]: preparing layers... -(22:05:29 2015-11-12)[nerv] info: create layer: recurrentL1 -(22:05:29 2015-11-12)[nerv] info: create layer: sigmoidL1 -(22:05:29 2015-11-12)[nerv] info: create layer: combinerL1 -(22:05:29 2015-11-12)[nerv] info: create layer: outputL -(22:05:29 2015-11-12)[nerv] info: create layer: softmaxL -(22:05:29 2015-11-12)[nerv] info: create layer: selectL1 -[SCHEDULER ITER6 LR1]: preparing layers end. -[SCHEDULER ITER6 LR1]: Generate and initing TNN ... - selectL1 0 -selectL1 recurrentL1 0 -recurrentL1 sigmoidL1 0 -sigmoidL1 combinerL1 0 -combinerL1 recurrentL1 1 -combinerL1 outputL 0 -outputL softmaxL 0 - softmaxL 0 -softmaxL 0 -recurrentL1 #dim_in: 2 #dim_out: 1 #i_conns_p: 2 #o_conns_p 1 -outputL #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -combinerL1 #dim_in: 1 #dim_out: 2 #i_conns_p: 1 #o_conns_p 2 -sigmoidL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -selectL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 0 #o_conns_p 1 -softmaxL #dim_in: 2 #dim_out: 1 #i_conns_p: 1 #o_conns_p 0 -TNN initing storage selectL1 -> recurrentL1 -TNN initing storage recurrentL1 -> sigmoidL1 -TNN initing storage sigmoidL1 -> combinerL1 -TNN initing storage combinerL1 -> recurrentL1 -TNN initing storage combinerL1 -> outputL -TNN initing storage outputL -> softmaxL -[SCHEDULER ITER6 LR1]: Initing TNN end. -===ITERATION 7 LR 1.000000=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER7 LR1]: 40092 words processed Thu Nov 12 22:05:45 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.247858. -[SCHEDULER ITER7 LR1]: 80099 words processed Thu Nov 12 22:06:01 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.240500. -[SCHEDULER ITER7 LR1]: 120004 words processed Thu Nov 12 22:06:17 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.233063. -[SCHEDULER ITER7 LR1]: 160114 words processed Thu Nov 12 22:06:33 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.233546. -[SCHEDULER ITER7 LR1]: 200066 words processed Thu Nov 12 22:06:49 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.233342. -[SCHEDULER ITER7 LR1]: 240045 words processed Thu Nov 12 22:07:05 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.228546. -[SCHEDULER ITER7 LR1]: 280057 words processed Thu Nov 12 22:07:21 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.226190. -[SCHEDULER ITER7 LR1]: 320106 words processed Thu Nov 12 22:07:37 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.222945. -[SCHEDULER ITER7 LR1]: 360024 words processed Thu Nov 12 22:07:53 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.221231. -[SCHEDULER ITER7 LR1]: 400089 words processed Thu Nov 12 22:08:09 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.219454. -[SCHEDULER ITER7 LR1]: 440067 words processed Thu Nov 12 22:08:25 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.219290. -[SCHEDULER ITER7 LR1]: 480051 words processed Thu Nov 12 22:08:41 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.219243. -[SCHEDULER ITER7 LR1]: 520140 words processed Thu Nov 12 22:08:57 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.218756. -[SCHEDULER ITER7 LR1]: 560132 words processed Thu Nov 12 22:09:13 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.218041. -[SCHEDULER ITER7 LR1]: 600118 words processed Thu Nov 12 22:09:29 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.218228. -[SCHEDULER ITER7 LR1]: 640090 words processed Thu Nov 12 22:09:45 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.217169. -[SCHEDULER ITER7 LR1]: 680075 words processed Thu Nov 12 22:10:01 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.216362. -[SCHEDULER ITER7 LR1]: 720043 words processed Thu Nov 12 22:10:17 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.216029. -[SCHEDULER ITER7 LR1]: 760012 words processed Thu Nov 12 22:10:33 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.215487. -[SCHEDULER ITER7 LR1]: 800113 words processed Thu Nov 12 22:10:49 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.215356. -[SCHEDULER ITER7 LR1]: 840089 words processed Thu Nov 12 22:11:05 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.214292. -[SCHEDULER ITER7 LR1]: 880052 words processed Thu Nov 12 22:11:21 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.214097. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER7 LR1]: Displaying result: -[SCHEDULER ITER7 LR1]: LMResult status of rnn: -[SCHEDULER ITER7 LR1]: Doing on /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf end. -[SCHEDULER ITER7 LR1]: shuffling training file -===PEEK ON TEST 7=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.test.txt.adds... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER7 LR1]: 40087 words processed Thu Nov 12 22:11:32 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.238654. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER7 LR1]: Displaying result: -[SCHEDULER ITER7 LR1]: LMResult status of rnn: -[SCHEDULER ITER7 LR1]: Doing on /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.test.txt.adds end. -===VALIDATION 7=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER7 LR1]: 40095 words processed Thu Nov 12 22:11:44 2015. - [SCHEDULER ITER7 LR1]: log prob per sample :-2.297186. -[LOG]LMSeqReader: file expires, closing. -[SCHEDULER ITER7 LR1]: Displaying result: -[SCHEDULER ITER7 LR1]: LMResult status of rnn: -[SCHEDULER ITER7 LR1]: Doing on /home/slhome/txh18/workspace/nerv/nerv/nerv/examples/lmptb/PTBdata/ptb.valid.txt.adds end. -[SCHEDULER ITER7 LR1]: saving net to file /home/slhome/txh18/workspace/nerv/play/dagL_test/params... - -[SCHEDULER ITER7 LR1]: preparing parameters... -reading chunk 0 from 0 -metadata: return {type="nerv.LinearTransParam",id="ltp_hh"} - -reading chunk 1 from 1073208 -metadata: return {type="nerv.BiasParam",id="bp_h"} - -reading chunk 2 from 1076866 -metadata: return {type="nerv.LinearTransParam",id="ltp_ho"} - -reading chunk 3 from 35689509 -metadata: return {type="nerv.BiasParam",id="bp_o"} - -reading chunk 4 from 35805861 -metadata: return {type="nerv.LinearTransParam",id="ltp_ih"} - -reading chunk 5 from 70341551 -[SCHEDULER ITER7 LR1]: preparing parameters end. -[SCHEDULER ITER7 LR1]: preparing layers... -(22:11:54 2015-11-12)[nerv] info: create layer: recurrentL1 -(22:11:54 2015-11-12)[nerv] info: create layer: sigmoidL1 -(22:11:54 2015-11-12)[nerv] info: create layer: combinerL1 -(22:11:54 2015-11-12)[nerv] info: create layer: outputL -(22:11:54 2015-11-12)[nerv] info: create layer: softmaxL -(22:11:54 2015-11-12)[nerv] info: create layer: selectL1 -[SCHEDULER ITER7 LR1]: preparing layers end. -[SCHEDULER ITER7 LR1]: Generate and initing TNN ... - selectL1 0 -selectL1 recurrentL1 0 -recurrentL1 sigmoidL1 0 -sigmoidL1 combinerL1 0 -combinerL1 recurrentL1 1 -combinerL1 outputL 0 -outputL softmaxL 0 - softmaxL 0 -softmaxL 0 -recurrentL1 #dim_in: 2 #dim_out: 1 #i_conns_p: 2 #o_conns_p 1 -outputL #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -combinerL1 #dim_in: 1 #dim_out: 2 #i_conns_p: 1 #o_conns_p 2 -sigmoidL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 1 #o_conns_p 1 -selectL1 #dim_in: 1 #dim_out: 1 #i_conns_p: 0 #o_conns_p 1 -softmaxL #dim_in: 2 #dim_out: 1 #i_conns_p: 1 #o_conns_p 0 -TNN initing storage selectL1 -> recurrentL1 -TNN initing storage recurrentL1 -> sigmoidL1 -TNN initing storage sigmoidL1 -> combinerL1 -TNN initing storage combinerL1 -> recurrentL1 -TNN initing storage combinerL1 -> outputL -TNN initing storage outputL -> softmaxL -[SCHEDULER ITER7 LR1]: Initing TNN end. -===ITERATION 8 LR 1.000000=== -[LOG]LMSeqReader: opening file /home/slhome/txh18/workspace/nerv/play/dagL_test/train_fn_shuf... -batch_size: 10 chunk_size 15 -[SCHEDULER ITER8 LR1]: 40092 words processed Thu Nov 12 22:12:10 2015. - [SCHEDULER ITER8 LR1]: log prob per sample :-2.217380. -[SCHEDULER ITER8 LR1]: 80099 words processed Thu Nov 12 22:12:26 2015. - [SCHEDULER ITER8 LR1]: log prob per sample :-2.210173. -[SCHEDULER ITER8 LR1]: 120004 words processed Thu