-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathlearning.py
More file actions
135 lines (101 loc) · 4.69 KB
/
learning.py
File metadata and controls
135 lines (101 loc) · 4.69 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
import torch
import numpy as np
from utils import Dataset, print_log
def fit_model(m, train_data, valid_data, loss_type='nll', with_done=False,
n_epochs=200, epoch_size=100, batch_size=16,
lr=1e-2, patience=10, log=False, logfile=None, min_int_ratio=0.0, threshold=1e-4):
# infer the device from the model
device = next(m.parameters()).device
if log:
print_log(f"loss_type: {loss_type}", logfile)
print_log(f"with_done: {with_done}", logfile)
print_log(f"n_epochs: {n_epochs}", logfile)
print_log(f"epoch_size: {epoch_size}", logfile)
print_log(f"batch_size: {batch_size}", logfile)
print_log(f"lr: {lr}", logfile)
print_log(f"patience: {patience}", logfile)
print_log(f"device: {device}", logfile)
print_log(f"min_int_ratio: {min_int_ratio}", logfile)
def compute_weights(data):
nint = np.sum([regime == 1 for regime, _ in data])
nobs = len(data) - nint
int_ratio = nint / (nint + nobs)
if int_ratio >= min_int_ratio:
weights = [1] * len(data)
else:
weights = [(1 - min_int_ratio) / nobs, min_int_ratio / nint] # obs, int
weights = [weights[int(regime)] for regime, _ in data]
return weights
train_weights = compute_weights(train_data)
valid_weights = compute_weights(valid_data)
# Build training and validation datasets and dataloaders
train_sampler = torch.utils.data.WeightedRandomSampler(train_weights, replacement=True, num_samples=epoch_size*batch_size)
train_dataset = Dataset(train_data)
train_loader = torch.utils.data.DataLoader(train_dataset, sampler=train_sampler, batch_size=batch_size)
valid_dataset = Dataset(list(zip(valid_data, valid_weights))) # to reweight the loss
valid_loader = torch.utils.data.DataLoader(valid_dataset, batch_size=batch_size)
# Adam Optimizer with learning rate lr
optimizer = torch.optim.Adam(m.parameters(), lr=lr)
# Scheduler. Reduce learning rate on plateau.
scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer, 'min', patience=patience, verbose=log, threshold=threshold)
# Early stopping
best_valid_loss = float("Inf")
best_parameters = m.state_dict().copy()
best_epoch = -1
# Start training loop
for epoch in range(n_epochs + 1):
# Set initial training loss as +inf
if epoch == 0:
train_loss = float("Inf")
else:
train_loss = 0
train_nsamples = 0
for batch in train_loader:
regime, episode = batch
regime = regime.to(device)
episode = [tensor.to(device) for tensor in episode]
batch_size = regime.shape[0]
if loss_type == 'em':
loss = m.loss_em(regime, episode, with_done=with_done).mean()
elif loss_type == 'nll':
loss = m.loss_nll(regime, episode, with_done=with_done).mean()
elif loss_type == 'elbo':
raise NotImplementedError()
else:
raise NotImplementedError()
optimizer.zero_grad()
loss.backward()
optimizer.step()
train_loss += loss.item() * batch_size
train_nsamples += batch_size
train_loss /= train_nsamples
# validation
valid_loss = 0
valid_nsamples = 0
for batch in valid_loader:
(regime, episode), weight = batch
regime = regime.to(device)
episode = [tensor.to(device) for tensor in episode]
weight = weight.to(device)
batch_size = regime.shape[0]
with torch.no_grad():
loss = m.loss_nll(regime, episode, with_done=with_done)
loss = (loss * weight).sum() # re-weighting the loss here
valid_loss += loss.item()
valid_nsamples += weight.sum().item()
valid_loss /= valid_nsamples
if log:
print_log(f"epoch {epoch:04d} / {n_epochs:04d} train loss={train_loss:0.3f} valid loss={valid_loss:0.3f}", logfile)
# check for best model
if valid_loss < (best_valid_loss * (1 - threshold)):
best_valid_loss = valid_loss
best_parameters = m.state_dict().copy()
best_epoch = epoch
# check for early stopping
if epoch > best_epoch + 2*patience:
if log:
print_log(f"{epoch-best_epoch} epochs without improvement, stopping.", logfile)
break
scheduler.step(valid_loss)
# restore best model
m.load_state_dict(best_parameters)