forked from gjy3035/NWPU-Crowd-Sample-Code
-
Notifications
You must be signed in to change notification settings - Fork 0
/
test.py
116 lines (93 loc) · 3.76 KB
/
test.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
from matplotlib import pyplot as plt
import matplotlib
import os
import random
import torch
from torch.autograd import Variable
import torchvision.transforms as standard_transforms
import misc.transforms as own_transforms
import pandas as pd
from models.CC import CrowdCounter
from config import cfg
from misc.utils import *
import scipy.io as sio
from PIL import Image, ImageOps
torch.cuda.set_device(0)
torch.backends.cudnn.benchmark = True
mean_std = ([0.446139603853, 0.409515678883, 0.395083993673], [0.288205742836, 0.278144598007, 0.283502370119])
img_transform = standard_transforms.Compose([
standard_transforms.ToTensor(),
standard_transforms.Normalize(*mean_std)
])
restore = standard_transforms.Compose([
own_transforms.DeNormalize(*mean_std),
standard_transforms.ToPILImage()
])
pil_to_tensor = standard_transforms.ToTensor()
LOG_PARA = 100.0
dataRoot = '../ProcessedData/Data.2019.11/NWPU/1204_min_576x768_mod16_2048'
#model_path = 'exp/12-06_15-03_NWPU_Res101_SFCN_1e-05/latest_state.pth'
model_path = 'exp/12-06_15-03_NWPU_Res101_SFCN_1e-05/all_ep_321_mae_90.7_mse_487.2_nae_0.375.pth'
def main():
txtpath = os.path.join(dataRoot, 'txt_list', 'test.txt')
with open(txtpath) as f:
lines = f.readlines()
test(lines, model_path)
def test(file_list, model_path):
net = CrowdCounter(cfg.GPU_ID, 'Res101_SFCN')
net.cuda()
#lastest_state = torch.load(model_path)
#net.load_state_dict(lastest_state['net'])
net.load_state_dict(torch.load(model_path))
net.eval()
gts = []
preds = []
record = open('submmited.txt', 'w+')
for infos in file_list:
filename = infos.split()[0]
imgname = os.path.join(dataRoot, 'img', filename + '.jpg')
img = Image.open(imgname)
if img.mode == 'L':
img = img.convert('RGB')
img = img_transform(img)[None, :, :, :]
with torch.no_grad():
img = Variable(img).cuda()
crop_imgs, crop_masks = [], []
b, c, h, w = img.shape
rh, rw = 576, 768
for i in range(0, h, rh):
gis, gie = max(min(h-rh, i), 0), min(h, i+rh)
for j in range(0, w, rw):
gjs, gje = max(min(w-rw, j), 0), min(w, j+rw)
crop_imgs.append(img[:, :, gis:gie, gjs:gje])
mask = torch.zeros(b, 1, h, w).cuda()
mask[:, :, gis:gie, gjs:gje].fill_(1.0)
crop_masks.append(mask)
crop_imgs, crop_masks = map(lambda x: torch.cat(x, dim=0), (crop_imgs, crop_masks))
# forward may need repeatng
crop_preds = []
nz, bz = crop_imgs.size(0), 1
for i in range(0, nz, bz):
gs, gt = i, min(nz, i+bz)
crop_pred = net.test_forward(crop_imgs[gs:gt])
crop_preds.append(crop_pred)
crop_preds = torch.cat(crop_preds, dim=0)
# splice them to the original size
idx = 0
pred_map = torch.zeros(b, 1, h, w).cuda()
for i in range(0, h, rh):
gis, gie = max(min(h-rh, i), 0), min(h, i+rh)
for j in range(0, w, rw):
gjs, gje = max(min(w-rw, j), 0), min(w, j+rw)
pred_map[:, :, gis:gie, gjs:gje] += crop_preds[idx]
idx += 1
# for the overlapping area, compute average value
mask = crop_masks.sum(dim=0).unsqueeze(0)
pred_map = pred_map / mask
pred_map = pred_map.cpu().data.numpy()[0,0,:,:]
pred = np.sum(pred_map) / LOG_PARA
print(f'{filename} {pred:.4f}', file=record)
print(f'{filename} {pred:.4f}')
record.close()
if __name__ == '__main__':
main()