forked from Element-Research/dpnn
-
Notifications
You must be signed in to change notification settings - Fork 0
/
VRClassReward.lua
93 lines (81 loc) · 3.33 KB
/
VRClassReward.lua
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
------------------------------------------------------------------------
--[[ VRClassReward ]]--
-- Variance reduced classification reinforcement criterion.
-- input : {class prediction, baseline reward}
-- Reward is 1 for success, Reward is 0 otherwise.
-- reward = scale*(Reward - baseline) where baseline is 2nd input element
-- Note : for RNNs with R = 1 for last step in sequence, encapsulate it
-- in nn.ModuleCriterion(VRClassReward, nn.SelectTable(-1))
------------------------------------------------------------------------
local VRClassReward, parent = torch.class("nn.VRClassReward", "nn.Criterion")
function VRClassReward:__init(module, scale, criterion)
parent.__init(self)
self.module = module -- so it can call module:reinforce(reward)
self.scale = scale or 1 -- scale of reward
self.criterion = criterion or nn.MSECriterion() -- baseline criterion
self.sizeAverage = true
self.gradInput = {torch.Tensor()}
end
function VRClassReward:updateOutput(input, target)
assert(torch.type(input) == 'table')
local input = self:toBatch(input[1], 1)
self._maxVal = self._maxVal or input.new()
self._maxIdx = self._maxIdx or torch.type(input) == 'torch.CudaTensor' and torch.CudaLongTensor() or torch.LongTensor()
-- max class value is class prediction
self._maxVal:max(self._maxIdx, input, 2)
-- reward = scale when correctly classified
local maxIdx = self._maxIdx
if torch.type(self._maxIdx) == 'torch.CudaLongTensor' then
self.__maxIdx = self.__maxIdx or torch.CudaTensor()
self.__maxIdx:resize(maxIdx:size()):copy(maxIdx)
maxIdx = self.__maxIdx
end
if torch.type(maxIdx) ~= torch.type(target) then
self._target = self._target or maxIdx.new()
self._target:resize(target:size()):copy(target)
target = self._target
end
self._reward = self._reward or maxIdx.new()
self._reward:eq(maxIdx, target)
self.reward = self.reward or input.new()
self.reward:resize(self._reward:size(1)):copy(self._reward)
self.reward:mul(self.scale)
-- loss = -sum(reward)
self.output = -self.reward:sum()
if self.sizeAverage then
self.output = self.output/input:size(1)
end
return self.output
end
function VRClassReward:updateGradInput(inputTable, target)
local input = self:toBatch(inputTable[1], 1)
local baseline = self:toBatch(inputTable[2], 1)
-- reduce variance of reward using baseline
self.vrReward = self.vrReward or self.reward.new()
self.vrReward:resizeAs(self.reward):copy(self.reward)
self.vrReward:add(-1, baseline)
if self.sizeAverage then
self.vrReward:div(input:size(1))
end
-- broadcast reward to modules
self.module:reinforce(self.vrReward)
-- zero gradInput (this criterion has no gradInput for class pred)
self.gradInput[1]:resizeAs(input):zero()
self.gradInput[1] = self:fromBatch(self.gradInput[1], 1)
-- learn the baseline reward
self.criterion:forward(baseline, self.reward)
self.gradInput[2] = self.criterion:backward(baseline, self.reward)
self.gradInput[2] = self:fromBatch(self.gradInput[2], 1)
return self.gradInput
end
function VRClassReward:type(type)
self._maxVal = nil
self._maxIdx = nil
self.__maxIdx = nil
self._target = nil
local module = self.module
self.module = nil
local ret = parent.type(self, type)
self.module = module
return ret
end