aboutsummaryrefslogtreecommitdiff
path: root/nerv/examples/test_nn_lib.lua
blob: 54448100a0dad88efda32c472eb2dadf25cbb910 (plain) (blame)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
require 'speech.init'
gconf = {lrate = 0.8, wcost = 1e-6, momentum = 0.9,
        cumat_type = nerv.CuMatrixFloat,
        mmat_type = nerv.MMatrixFloat,
        batch_size = 256}

param_repo = nerv.ParamRepo({"converted.nerv", "global_transf.nerv"})
sublayer_repo = nerv.LayerRepo(
    {
        -- global transf
        ["nerv.BiasLayer"] =
        {
            blayer1 = {{bias = "bias1"}, {dim_in = {429}, dim_out = {429}}},
            blayer2 = {{bias = "bias2"}, {dim_in = {429}, dim_out = {429}}}
        },
        ["nerv.WindowLayer"] =
        {
            wlayer1 = {{window = "window1"}, {dim_in = {429}, dim_out = {429}}},
            wlayer2 = {{window = "window2"}, {dim_in = {429}, dim_out = {429}}}
        },
        -- biased linearity
        ["nerv.AffineLayer"] =
        {
            affine0 = {{ltp = "affine0_ltp", bp = "affine0_bp"},
                        {dim_in = {429}, dim_out = {2048}}},
            affine1 = {{ltp = "affine1_ltp", bp = "affine1_bp"},
                        {dim_in = {2048}, dim_out = {2048}}},
            affine2 = {{ltp = "affine2_ltp", bp = "affine2_bp"},
                        {dim_in = {2048}, dim_out = {2048}}},
            affine3 = {{ltp = "affine3_ltp", bp = "affine3_bp"},
                        {dim_in = {2048}, dim_out = {2048}}},
            affine4 = {{ltp = "affine4_ltp", bp = "affine4_bp"},
                        {dim_in = {2048}, dim_out = {2048}}},
            affine5 = {{ltp = "affine5_ltp", bp = "affine5_bp"},
                        {dim_in = {2048}, dim_out = {2048}}},
            affine6 = {{ltp = "affine6_ltp", bp = "affine6_bp"},
                        {dim_in = {2048}, dim_out = {2048}}},
            affine7 = {{ltp = "affine7_ltp", bp = "affine7_bp"},
                        {dim_in = {2048}, dim_out = {3001}}}
        },
        ["nerv.SigmoidLayer"] =
        {
            sigmoid0 = {{}, {dim_in = {2048}, dim_out = {2048}}},
            sigmoid1 = {{}, {dim_in = {2048}, dim_out = {2048}}},
            sigmoid2 = {{}, {dim_in = {2048}, dim_out = {2048}}},
            sigmoid3 = {{}, {dim_in = {2048}, dim_out = {2048}}},
            sigmoid4 = {{}, {dim_in = {2048}, dim_out = {2048}}},
            sigmoid5 = {{}, {dim_in = {2048}, dim_out = {2048}}},
            sigmoid6 = {{}, {dim_in = {2048}, dim_out = {2048}}}
        },
        ["nerv.SoftmaxCELayer"] =
        {
            softmax_ce0 = {{}, {dim_in = {3001, 1}, dim_out = {}, compressed = true}}
        }
    }, param_repo, gconf)

layer_repo = nerv.LayerRepo(
    {
        ["nerv.DAGLayer"] =
        {
            global_transf = {{}, {
                dim_in = {429}, dim_out = {429},
                sub_layers = sublayer_repo,
                connections = {
                    ["<input>[1]"] = "blayer1[1]",
                    ["blayer1[1]"] = "wlayer1[1]",
                    ["wlayer1[1]"] = "blayer2[1]",
                    ["blayer2[1]"] = "wlayer2[1]",
                    ["wlayer2[1]"] = "<output>[1]"
                }
            }},
            main = {{}, {
                dim_in = {429, 1}, dim_out = {},
                sub_layers = sublayer_repo,
                connections = {
                    ["<input>[1]"] = "affine0[1]",
                    ["affine0[1]"] = "sigmoid0[1]",
                    ["sigmoid0[1]"] = "affine1[1]",
                    ["affine1[1]"] = "sigmoid1[1]",
                    ["sigmoid1[1]"] = "affine2[1]",
                    ["affine2[1]"] = "sigmoid2[1]",
                    ["sigmoid2[1]"] = "affine3[1]",
                    ["affine3[1]"] = "sigmoid3[1]",
                    ["sigmoid3[1]"] = "affine4[1]",
                    ["affine4[1]"] = "sigmoid4[1]",
                    ["sigmoid4[1]"] = "affine5[1]",
                    ["affine5[1]"] = "sigmoid5[1]",
                    ["sigmoid5[1]"] = "affine6[1]",
                    ["affine6[1]"] = "sigmoid6[1]",
                    ["sigmoid6[1]"] = "affine7[1]",
                    ["affine7[1]"] = "softmax_ce0[1]",
                    ["<input>[2]"] = "softmax_ce0[2]"
                }
            }}
        }
    }, param_repo, gconf)

tnet_reader = nerv.TNetReader(gconf,
    {
        id = "main_scp",
        scp_file = "/slfs1/users/mfy43/swb_ivec/train_bp.scp",
--        scp_file = "t.scp",
        conf_file = "/slfs1/users/mfy43/swb_ivec/plp_0_d_a.conf",
        frm_ext = 5,
        mlfs = {
            ref = {
                file = "/slfs1/users/mfy43/swb_ivec/ref.mlf",
                format = "map",
                format_arg = "/slfs1/users/mfy43/swb_ivec/dict",
                dir = "*/",
                ext = "lab"
            }
        },
        global_transf = layer_repo:get_layer("global_transf")
    })

buffer = nerv.SGDBuffer(gconf,
    {
        buffer_size = 81920,
        randomize = true,
        readers = {
            { reader = tnet_reader,
              data = {main_scp = 429, ref = 1}}
        }
    })

sm = sublayer_repo:get_layer("softmax_ce0")
main = layer_repo:get_layer("main")
main:init(gconf.batch_size)
gconf.cnt = 0
-- data = buffer:get_data()
-- input = {data.main_scp, data.ref}
-- while true do
for data in buffer.get_data, buffer do
--    if gconf.cnt == 100 then break end
--    gconf.cnt = gconf.cnt + 1

    input = {data.main_scp, data.ref}
    output = {}
    err_input = {}
    err_output = {input[1]:create()}
    
    main:propagate(input, output)
    main:back_propagate(err_output, err_input, input, output)
    main:update(err_input, input, output)

--    nerv.printf("cross entropy: %.8f\n", sm.total_ce)
--    nerv.printf("correct: %d\n", sm.total_correct)
--    nerv.printf("frames: %d\n", sm.total_frames)
--    nerv.printf("err/frm: %.8f\n", sm.total_ce / sm.total_frames)
--    nerv.printf("accuracy: %.8f\n", sm.total_correct / sm.total_frames)
    collectgarbage("collect")
end
nerv.printf("cross entropy: %.8f\n", sm.total_ce)
nerv.printf("correct: %d\n", sm.total_correct)
nerv.printf("accuracy: %.3f%%\n", sm.total_correct / sm.total_frames * 100)
nerv.printf("writing back...\n")
cf = nerv.ChunkFile("output.nerv", "w")
for i, p in ipairs(main:get_params()) do
    print(p)
    cf:write_chunk(p)
end
cf:close()
nerv.Matrix.print_profile()