-
Notifications
You must be signed in to change notification settings - Fork 19
Expand file tree
/
Copy pathtest_custom.py
More file actions
97 lines (72 loc) · 3.37 KB
/
test_custom.py
File metadata and controls
97 lines (72 loc) · 3.37 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
import os
import json
import argparse
import warnings
import torch
import torch.nn as nn
import torch.optim as optim
from torch.utils.data import DataLoader
from utils import *
from loader import EEGDataLoader
from models.main_model import MainModel
class OneFoldTrainer:
def __init__(self, args, fold, config):
self.args = args
self.fold = fold
self.cfg = config
self.ds_cfg = config['dataset']
self.fp_cfg = config['feature_pyramid']
self.tp_cfg = config['training_params']
self.es_cfg = self.tp_cfg['early_stopping']
self.device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
print('[INFO] Config name: {}'.format(config['name']))
self.train_iter = 0
self.model = self.build_model()
self.ckpt_path = os.path.join('checkpoints', config['name'])
self.ckpt_name = 'ckpt_fold-{0:02d}.pth'.format(self.fold)
self.early_stopping = EarlyStopping(patience=self.es_cfg['patience'], verbose=True, ckpt_path=self.ckpt_path, ckpt_name=self.ckpt_name, mode=self.es_cfg['mode'])
def build_model(self):
model = MainModel(self.cfg)
print('[INFO] Number of params of model: ', sum(p.numel() for p in model.parameters() if p.requires_grad))
model = torch.nn.DataParallel(model, device_ids=list(range(len(self.args.gpu.split(",")))))
load_path = os.path.join('checkpoints', self.cfg['name'], 'ckpt_fold-{0:02d}.pth'.format(self.fold))
model.load_state_dict(torch.load(load_path), strict=False)
print('[INFO] Model loaded')
model.to(self.device)
print('[INFO] Model prepared, Device used: {} GPU:{}'.format(self.device, self.args.gpu))
return model
@torch.no_grad()
def evaluate(self, input_data):
inputs = torch.tensor(input_data, dtype=torch.float32).to(self.device)
self.model.eval()
outputs = self.model(inputs)
outputs_sum = torch.zeros_like(outputs[0])
for output in outputs:
outputs_sum += output
predicted = torch.argmax(outputs_sum, 1)
print('Predicted: ', predicted)
def run(self):
self.model.load_state_dict(torch.load(os.path.join(self.ckpt_path, self.ckpt_name)))
input_data = np.random.rand(1, 1, 30000)
self.evaluate(input_data)
print('')
def main():
warnings.filterwarnings("ignore", category=DeprecationWarning)
warnings.filterwarnings("ignore", category=UserWarning)
parser = argparse.ArgumentParser(formatter_class=argparse.ArgumentDefaultsHelpFormatter)
parser.add_argument('--seed', type=int, default=42, help='random seed')
parser.add_argument('--gpu', type=str, default="0", help='gpu id')
parser.add_argument('--config', type=str, help='config file path')
parser.add_argument('--fold', type=int, default=1, help='fold to load checkpoint')
args = parser.parse_args()
os.environ["CUDA_DEVICE_ORDER"] = "PCI_BUS_ID"
os.environ["CUDA_VISIBLE_DEVICES"] = args.gpu
# For reproducibility
# set_random_seed(args.seed, use_cuda=True)
with open(args.config) as config_file:
config = json.load(config_file)
config['name'] = os.path.basename(args.config).replace('.json', '')
trainer = OneFoldTrainer(args, args.fold, config)
trainer.run()
if __name__ == "__main__":
main()