forked from chengtan9907/Co-learning
-
Notifications
You must be signed in to change notification settings - Fork 0
/
JoCoR.py
118 lines (95 loc) · 4.2 KB
/
JoCoR.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
import torch
import torch.nn.functional as F
from torch.autograd import Variable
import numpy as np
from utils import get_model
from tqdm import tqdm
from losses import loss_jocor
device = torch.device('cuda') if torch.cuda.is_available() else torch.device('cpu')
class JoCoR:
def __init__(
self,
config: dict = None,
input_channel: int = 3,
num_classes: int = 10,
):
self.lr = config['lr']
if config['forget_rate'] is None:
if config['noise_type'] == 'asym':
forget_rate = config['percent'] / 2
else:
forget_rate = config['percent']
else:
forget_rate = config['forget_rate']
# Adjust learning rate and betas for Adam Optimizer
mom1 = 0.9
mom2 = 0.1
self.alpha_plan = [self.lr] * config['epochs']
self.beta1_plan = [mom1] * config['epochs']
for i in range(config['epoch_decay_start'], config['epochs']):
self.alpha_plan[i] = float(config['epochs'] - i) / (config['epochs'] - config['epoch_decay_start']) * self.lr
self.beta1_plan[i] = mom2
# define drop rate schedule
self.rate_schedule = np.ones(config['epochs']) * forget_rate
self.rate_schedule[:config['num_gradual']] = np.linspace(0, forget_rate ** config['exponent'], config['num_gradual'])
self.device = device
self.co_lambda = config['co_lambda']
self.epochs = config['epochs']
# model
self.model1 = get_model(config['model1_type'], input_channel, num_classes, device)
self.model2 = get_model(config['model2_type'], input_channel, num_classes, device)
self.optimizer = torch.optim.Adam(list(self.model1.parameters()) + list(self.model2.parameters()),
lr=self.lr)
self.loss_fn = loss_jocor
self.adjust_lr = config['adjust_lr']
def evaluate(self, test_loader):
print('Evaluating ...')
self.model1.eval() # Change model to 'eval' mode.
self.model2.eval() # Change model to 'eval' mode
correct1 = 0
total1 = 0
for images, labels in test_loader:
images = Variable(images).to(self.device)
logits1 = self.model1(images)
outputs1 = F.softmax(logits1, dim=1)
_, pred1 = torch.max(outputs1.data, 1)
total1 += labels.size(0)
correct1 += (pred1.cpu() == labels).sum()
correct2 = 0
total2 = 0
for images, labels in test_loader:
images = Variable(images).to(self.device)
logits2 = self.model2(images)
outputs2 = F.softmax(logits2, dim=1)
_, pred2 = torch.max(outputs2.data, 1)
total2 += labels.size(0)
correct2 += (pred2.cpu() == labels).sum()
acc1 = 100 * float(correct1) / float(total1)
acc2 = 100 * float(correct2) / float(total2)
return acc1, acc2
def train(self, train_loader, epoch):
print('Training ...')
self.model1.train()
self.model2.train()
if self.adjust_lr == 1:
self.adjust_learning_rate(self.optimizer, epoch)
pbar = tqdm(train_loader)
for (images, labels) in pbar:
images = Variable(images).to(self.device)
labels = Variable(labels).to(self.device)
# Forward + Backward + Optimize
logits1 = self.model1(images)
logits2 = self.model2(images)
loss_1, loss_2 = self.loss_fn(logits1, logits2, labels,
self.rate_schedule[epoch],
self.co_lambda)
self.optimizer.zero_grad()
loss_1.backward()
self.optimizer.step()
pbar.set_description(
'Epoch [%d/%d], Loss1: %.4f, Loss2: %.4f'
% (epoch + 1, self.epochs, loss_1.data.item(), loss_2.data.item()))
def adjust_learning_rate(self, optimizer, epoch):
for param_group in optimizer.param_groups:
param_group['lr'] = self.alpha_plan[epoch]
param_group['betas'] = (self.beta1_plan[epoch], 0.999) # Only change beta1