summaryrefslogtreecommitdiff
path: root/embedding_example/setup_nerv.lua
blob: e33a1e79b0b735d4a396bc5f86ca85734e16c0da (plain) (blame)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
package.path="/home/slhome/mfy43/.luarocks/share/lua/5.1/?.lua;/home/slhome/mfy43/.luarocks/share/lua/5.1/?/init.lua;/home/slhome/mfy43/nerv/install/share/lua/5.1/?.lua;/home/slhome/mfy43/nerv/install/share/lua/5.1/?/init.lua;"..package.path
package.cpath="/home/slhome/mfy43/.luarocks/lib/lua/5.1/?.so;/home/slhome/mfy43/nerv/install/lib/lua/5.1/?.so;"..package.cpath
local k,l,_=pcall(require,"luarocks.loader") _=k and l.add_context("nerv","scm-1")

local args = {...}
require 'nerv'
dofile(args[1])
local param_repo = nerv.ParamRepo()
param_repo:import(gconf.initialized_param, nil, gconf)
local sublayer_repo = make_sublayer_repo(param_repo)
local layer_repo = make_layer_repo(sublayer_repo, param_repo)
local network = get_network(layer_repo)
local batch_size = 1
network:init(batch_size)
function propagator(input, output)
    local gpu_input = nerv.CuMatrixFloat(input:nrow(), input:ncol())
    local gpu_output = nerv.CuMatrixFloat(output:nrow(), output:ncol())
    gpu_input:copy_fromh(input)
    print(gpu_input)
    network:propagate({gpu_input}, {gpu_output})
    gpu_output:copy_toh(output)
    print(output)
    -- collect garbage in-time to save GPU memory
    collectgarbage("collect")
end
return network.dim_in[1], network.dim_out[1], propagator