2020-12-18 16:10:44 +00:00
|
|
|
from torch.cuda.amp import GradScaler
|
2020-10-22 19:58:05 +00:00
|
|
|
|
2020-08-22 14:24:34 +00:00
|
|
|
from utils.loss_accumulator import LossAccumulator
|
|
|
|
from torch.nn import Module
|
|
|
|
import logging
|
2020-12-18 16:18:34 +00:00
|
|
|
from trainer.losses import create_loss
|
2020-08-22 14:24:34 +00:00
|
|
|
import torch
|
|
|
|
from collections import OrderedDict
|
2020-12-30 03:58:02 +00:00
|
|
|
from trainer.inject import create_injector
|
2020-09-20 03:47:34 +00:00
|
|
|
from utils.util import recursively_detach
|
2020-08-12 14:45:23 +00:00
|
|
|
|
2020-08-22 14:24:34 +00:00
|
|
|
logger = logging.getLogger('base')
|
2020-08-12 14:45:23 +00:00
|
|
|
|
|
|
|
|
2020-08-22 14:24:34 +00:00
|
|
|
# Defines the expected API for a single training step
|
|
|
|
class ConfigurableStep(Module):
|
|
|
|
|
|
|
|
def __init__(self, opt_step, env):
|
|
|
|
super(ConfigurableStep, self).__init__()
|
|
|
|
|
|
|
|
self.step_opt = opt_step
|
|
|
|
self.env = env
|
|
|
|
self.opt = env['opt']
|
|
|
|
self.gen_outputs = opt_step['generator_outputs']
|
|
|
|
self.loss_accumulator = LossAccumulator()
|
2020-08-23 23:22:34 +00:00
|
|
|
self.optimizers = None
|
2020-10-22 19:58:05 +00:00
|
|
|
self.scaler = GradScaler(enabled=self.opt['fp16'])
|
2020-10-22 20:39:19 +00:00
|
|
|
self.grads_generated = False
|
2020-11-05 17:04:17 +00:00
|
|
|
self.min_total_loss = opt_step['min_total_loss'] if 'min_total_loss' in opt_step.keys() else -999999999
|
2020-08-22 14:24:34 +00:00
|
|
|
|
|
|
|
self.injectors = []
|
|
|
|
if 'injectors' in self.step_opt.keys():
|
2020-09-29 22:59:44 +00:00
|
|
|
injector_names = []
|
2020-08-22 14:24:34 +00:00
|
|
|
for inj_name, injector in self.step_opt['injectors'].items():
|
2020-09-29 22:59:44 +00:00
|
|
|
assert inj_name not in injector_names # Repeated names are always an error case.
|
|
|
|
injector_names.append(inj_name)
|
2020-08-22 14:24:34 +00:00
|
|
|
self.injectors.append(create_injector(injector, env))
|
|
|
|
|
|
|
|
losses = []
|
|
|
|
self.weights = {}
|
2020-08-31 15:41:48 +00:00
|
|
|
if 'losses' in self.step_opt.keys():
|
|
|
|
for loss_name, loss in self.step_opt['losses'].items():
|
2020-09-29 22:59:44 +00:00
|
|
|
assert loss_name not in self.weights.keys() # Repeated names are always an error case.
|
2020-09-29 04:09:42 +00:00
|
|
|
losses.append((loss_name, create_loss(loss, env)))
|
2020-08-31 15:41:48 +00:00
|
|
|
self.weights[loss_name] = loss['weight']
|
2020-08-22 14:24:34 +00:00
|
|
|
self.losses = OrderedDict(losses)
|
|
|
|
|
2020-09-12 04:57:06 +00:00
|
|
|
def get_network_for_name(self, name):
|
|
|
|
return self.env['generators'][name] if name in self.env['generators'].keys() \
|
|
|
|
else self.env['discriminators'][name]
|
|
|
|
|
2020-08-22 14:24:34 +00:00
|
|
|
# Subclasses should override this to define individual optimizers. They should all go into self.optimizers.
|
|
|
|
# This default implementation defines a single optimizer for all Generator parameters.
|
2020-08-23 23:22:34 +00:00
|
|
|
# Must be called after networks are initialized and wrapped.
|
2020-08-22 14:24:34 +00:00
|
|
|
def define_optimizers(self):
|
2020-09-12 04:57:06 +00:00
|
|
|
training = self.step_opt['training']
|
2020-12-07 19:50:57 +00:00
|
|
|
training_net = self.get_network_for_name(training)
|
|
|
|
# When only training one network, optimizer params can just embedded in the step params.
|
|
|
|
if 'optimizer_params' not in self.step_opt.keys():
|
|
|
|
opt_configs = [self.step_opt]
|
2020-09-12 04:57:06 +00:00
|
|
|
else:
|
2020-12-07 19:50:57 +00:00
|
|
|
opt_configs = [self.step_opt['optimizer_params']]
|
|
|
|
nets = [training_net]
|
|
|
|
training = [training]
|
2020-09-12 04:57:06 +00:00
|
|
|
self.optimizers = []
|
2020-11-05 01:07:48 +00:00
|
|
|
for net_name, net, opt_config in zip(training, nets, opt_configs):
|
2021-01-02 22:10:06 +00:00
|
|
|
# Configs can organize parameters by-group and specify different learning rates for each group. This only
|
|
|
|
# works in the model specifically annotates which parameters belong in which group using PARAM_GROUP.
|
|
|
|
optim_params = {'default': {'params': [], 'lr': opt_config['lr']}}
|
|
|
|
if 'param_groups' in opt_config.keys():
|
|
|
|
for k, pg in opt_config['param_groups'].items():
|
|
|
|
optim_params[k] = {'params': [], 'lr': pg['lr']}
|
|
|
|
|
2020-09-12 04:57:06 +00:00
|
|
|
for k, v in net.named_parameters(): # can optimize for a part of the model
|
2020-12-24 03:33:43 +00:00
|
|
|
# Make some inference about these parameters, which can be used by some optimizers to treat certain
|
|
|
|
# parameters differently. For example, it is considered good practice to not do weight decay on
|
|
|
|
# BN & bias parameters. TODO: process the module tree instead of the parameter tree to accomplish the
|
|
|
|
# same thing, but in a more effective way.
|
|
|
|
if k.endswith(".bias"):
|
|
|
|
v.is_bias = True
|
|
|
|
if k.endswith(".weight"):
|
|
|
|
v.is_weight = True
|
|
|
|
if ".bn" in k or '.batchnorm' in k or '.bnorm' in k:
|
|
|
|
v.is_bn = True
|
2021-01-02 22:10:06 +00:00
|
|
|
# Some models can specify some parameters to be in different groups.
|
|
|
|
param_group = "default"
|
|
|
|
if hasattr(v, 'PARAM_GROUP'):
|
|
|
|
if v.PARAM_GROUP in optim_params.keys():
|
|
|
|
param_group = v.PARAM_GROUP
|
|
|
|
else:
|
|
|
|
logger.warning(f'Model specifies a custom param group {v.PARAM_GROUP} which is not configured. '
|
|
|
|
f'The same LR will be used for all parameters.')
|
|
|
|
|
2020-09-12 04:57:06 +00:00
|
|
|
if v.requires_grad:
|
2021-01-02 22:10:06 +00:00
|
|
|
optim_params[param_group]['params'].append(v)
|
2020-09-12 04:57:06 +00:00
|
|
|
else:
|
|
|
|
if self.env['rank'] <= 0:
|
|
|
|
logger.warning('Params [{:s}] will not optimize.'.format(k))
|
|
|
|
|
|
|
|
if 'optimizer' not in self.step_opt.keys() or self.step_opt['optimizer'] == 'adam':
|
2021-01-02 22:10:06 +00:00
|
|
|
opt = torch.optim.Adam(list(optim_params.values()),
|
2020-09-12 04:57:06 +00:00
|
|
|
weight_decay=opt_config['weight_decay'],
|
|
|
|
betas=(opt_config['beta1'], opt_config['beta2']))
|
2020-12-24 03:33:43 +00:00
|
|
|
elif self.step_opt['optimizer'] == 'lars':
|
|
|
|
from trainer.optimizers.larc import LARC
|
|
|
|
from trainer.optimizers.sgd import SGDNoBiasMomentum
|
2021-01-03 05:22:34 +00:00
|
|
|
optSGD = SGDNoBiasMomentum(list(optim_params.values()), lr=opt_config['lr'], momentum=opt_config['momentum'],
|
2020-12-24 03:33:43 +00:00
|
|
|
weight_decay=opt_config['weight_decay'])
|
|
|
|
opt = LARC(optSGD, trust_coefficient=opt_config['lars_coefficient'])
|
2021-01-12 03:09:16 +00:00
|
|
|
elif self.step_opt['optimizer'] == 'sgd':
|
|
|
|
from torch.optim import SGD
|
|
|
|
opt = SGD(list(optim_params.values()), lr=opt_config['lr'], momentum=opt_config['momentum'], weight_decay=opt_config['weight_decay'])
|
2020-09-22 23:03:22 +00:00
|
|
|
opt._config = opt_config # This is a bit seedy, but we will need these configs later.
|
2020-11-05 01:07:48 +00:00
|
|
|
opt._config['network'] = net_name
|
2020-09-12 04:57:06 +00:00
|
|
|
self.optimizers.append(opt)
|
2020-08-12 14:45:23 +00:00
|
|
|
|
|
|
|
# Returns all optimizers used in this step.
|
|
|
|
def get_optimizers(self):
|
2020-08-22 14:24:34 +00:00
|
|
|
assert self.optimizers is not None
|
|
|
|
return self.optimizers
|
2020-08-12 14:45:23 +00:00
|
|
|
|
|
|
|
# Returns optimizers which are opting in for default LR scheduling.
|
|
|
|
def get_optimizers_with_default_scheduler(self):
|
2020-08-22 14:24:34 +00:00
|
|
|
assert self.optimizers is not None
|
|
|
|
return self.optimizers
|
2020-08-12 14:45:23 +00:00
|
|
|
|
|
|
|
# Returns the names of the networks this step will train. Other networks will be frozen.
|
|
|
|
def get_networks_trained(self):
|
2020-09-12 04:57:06 +00:00
|
|
|
if isinstance(self.step_opt['training'], list):
|
|
|
|
return self.step_opt['training']
|
|
|
|
else:
|
|
|
|
return [self.step_opt['training']]
|
2020-08-22 14:24:34 +00:00
|
|
|
|
2020-09-29 04:09:42 +00:00
|
|
|
def get_training_network_name(self):
|
|
|
|
if isinstance(self.step_opt['training'], list):
|
|
|
|
return self.step_opt['training'][0]
|
|
|
|
else:
|
|
|
|
return self.step_opt['training']
|
|
|
|
|
2020-08-22 14:24:34 +00:00
|
|
|
# Performs all forward and backward passes for this step given an input state. All input states are lists of
|
|
|
|
# chunked tensors. Use grad_accum_step to dereference these steps. Should return a dict of tensors that later
|
|
|
|
# steps might use. These tensors are automatically detached and accumulated into chunks.
|
2020-09-05 13:59:45 +00:00
|
|
|
def do_forward_backward(self, state, grad_accum_step, amp_loss_id, train=True):
|
2020-12-30 16:51:59 +00:00
|
|
|
local_state = {} # <-- Will store the entire local state to be passed to injectors & losses.
|
|
|
|
new_state = {} # <-- Will store state values created by this step for returning to ExtensibleTrainer.
|
2020-08-22 14:24:34 +00:00
|
|
|
for k, v in state.items():
|
|
|
|
local_state[k] = v[grad_accum_step]
|
2020-09-18 03:05:32 +00:00
|
|
|
local_state['train_nets'] = str(self.get_networks_trained())
|
2020-08-12 14:45:23 +00:00
|
|
|
|
2020-10-22 19:58:05 +00:00
|
|
|
# Some losses compute backward() internally. Accommodate this by stashing the amp_loss_id in env.
|
2020-09-25 22:38:23 +00:00
|
|
|
self.env['amp_loss_id'] = amp_loss_id
|
|
|
|
self.env['current_step_optimizers'] = self.optimizers
|
2020-09-27 03:25:32 +00:00
|
|
|
self.env['training'] = train
|
2020-09-25 22:38:23 +00:00
|
|
|
|
2020-10-28 02:58:03 +00:00
|
|
|
# Inject in any extra dependencies.
|
|
|
|
for inj in self.injectors:
|
|
|
|
# Don't do injections tagged with eval unless we are not in train mode.
|
|
|
|
if train and 'eval' in inj.opt.keys() and inj.opt['eval']:
|
|
|
|
continue
|
|
|
|
# Likewise, don't do injections tagged with train unless we are not in eval.
|
|
|
|
if not train and 'train' in inj.opt.keys() and inj.opt['train']:
|
|
|
|
continue
|
|
|
|
# Don't do injections tagged with 'after' or 'before' when we are out of spec.
|
|
|
|
if 'after' in inj.opt.keys() and self.env['step'] < inj.opt['after'] or \
|
2020-11-20 06:47:24 +00:00
|
|
|
'before' in inj.opt.keys() and self.env['step'] > inj.opt['before'] or \
|
2020-11-24 16:24:02 +00:00
|
|
|
'every' in inj.opt.keys() and self.env['step'] % inj.opt['every'] != 0:
|
2020-10-28 02:58:03 +00:00
|
|
|
continue
|
|
|
|
injected = inj(local_state)
|
|
|
|
local_state.update(injected)
|
|
|
|
new_state.update(injected)
|
|
|
|
|
|
|
|
if train and len(self.losses) > 0:
|
|
|
|
# Finally, compute the losses.
|
|
|
|
total_loss = 0
|
|
|
|
for loss_name, loss in self.losses.items():
|
|
|
|
# Some losses only activate after a set number of steps. For example, proto-discriminator losses can
|
|
|
|
# be very disruptive to a generator.
|
2020-11-03 15:09:58 +00:00
|
|
|
if 'after' in loss.opt.keys() and loss.opt['after'] > self.env['step'] or \
|
2020-11-12 22:42:05 +00:00
|
|
|
'before' in loss.opt.keys() and self.env['step'] > loss.opt['before'] or \
|
|
|
|
'every' in loss.opt.keys() and self.env['step'] % loss.opt['every'] != 0:
|
2020-09-09 02:36:27 +00:00
|
|
|
continue
|
2020-12-30 16:51:59 +00:00
|
|
|
if loss.is_stateful():
|
|
|
|
l, lstate = loss(self.get_network_for_name(self.step_opt['training']), local_state)
|
|
|
|
local_state.update(lstate)
|
|
|
|
new_state.update(lstate)
|
|
|
|
else:
|
|
|
|
l = loss(self.get_network_for_name(self.step_opt['training']), local_state)
|
2020-10-28 02:58:03 +00:00
|
|
|
total_loss += l * self.weights[loss_name]
|
|
|
|
# Record metrics.
|
|
|
|
if isinstance(l, torch.Tensor):
|
|
|
|
self.loss_accumulator.add_loss(loss_name, l)
|
|
|
|
for n, v in loss.extra_metrics():
|
|
|
|
self.loss_accumulator.add_loss("%s_%s" % (loss_name, n), v)
|
|
|
|
loss.clear_metrics()
|
|
|
|
|
|
|
|
# In some cases, the loss could not be set (e.g. all losses have 'after')
|
|
|
|
if isinstance(total_loss, torch.Tensor):
|
|
|
|
self.loss_accumulator.add_loss("%s_total" % (self.get_training_network_name(),), total_loss)
|
2020-10-28 21:46:59 +00:00
|
|
|
reset_required = total_loss < self.min_total_loss
|
|
|
|
|
2020-10-28 02:58:03 +00:00
|
|
|
# Scale the loss down by the accumulation factor.
|
|
|
|
total_loss = total_loss / self.env['mega_batch_factor']
|
|
|
|
|
|
|
|
# Get dem grads!
|
|
|
|
self.scaler.scale(total_loss).backward()
|
2020-10-28 21:46:59 +00:00
|
|
|
|
|
|
|
if reset_required:
|
|
|
|
# You might be scratching your head at this. Why would you zero grad as opposed to not doing a
|
|
|
|
# backwards? Because DDP uses the backward() pass as a synchronization point and there is not a good
|
|
|
|
# way to simply bypass backward. If you want a more efficient way to specify a min_loss, use or
|
|
|
|
# implement it at the loss level.
|
2020-12-07 19:50:57 +00:00
|
|
|
self.get_network_for_name(self.step_opt['training']).zero_grad()
|
2020-10-28 21:46:59 +00:00
|
|
|
self.loss_accumulator.increment_metric("%s_skipped_steps" % (self.get_training_network_name(),))
|
|
|
|
|
2020-10-28 02:58:03 +00:00
|
|
|
self.grads_generated = True
|
2020-08-22 14:24:34 +00:00
|
|
|
|
2020-08-23 23:22:34 +00:00
|
|
|
# Detach all state variables. Within the step, gradients can flow. Once these variables leave the step
|
|
|
|
# we must release the gradients.
|
2020-09-20 03:47:34 +00:00
|
|
|
new_state = recursively_detach(new_state)
|
2020-08-22 14:24:34 +00:00
|
|
|
return new_state
|
|
|
|
|
|
|
|
# Performs the optimizer step after all gradient accumulation is completed. Default implementation simply steps()
|
|
|
|
# all self.optimizers.
|
2020-11-05 01:07:48 +00:00
|
|
|
def do_step(self, step):
|
2020-10-22 20:39:19 +00:00
|
|
|
if not self.grads_generated:
|
|
|
|
return
|
|
|
|
self.grads_generated = False
|
2020-08-22 14:24:34 +00:00
|
|
|
for opt in self.optimizers:
|
2020-09-22 23:03:22 +00:00
|
|
|
# Optimizers can be opted out in the early stages of training.
|
|
|
|
after = opt._config['after'] if 'after' in opt._config.keys() else 0
|
2020-11-05 01:07:48 +00:00
|
|
|
after_network = self.opt['networks'][opt._config['network']]['after'] if 'after' in self.opt['networks'][opt._config['network']].keys() else 0
|
|
|
|
after = max(after, after_network)
|
2020-09-22 23:03:22 +00:00
|
|
|
if self.env['step'] < after:
|
|
|
|
continue
|
|
|
|
before = opt._config['before'] if 'before' in opt._config.keys() else -1
|
|
|
|
if before != -1 and self.env['step'] > before:
|
|
|
|
continue
|
2020-10-22 19:58:05 +00:00
|
|
|
self.scaler.step(opt)
|
|
|
|
self.scaler.update()
|
2020-08-22 14:24:34 +00:00
|
|
|
|
|
|
|
def get_metrics(self):
|
2020-08-22 19:08:33 +00:00
|
|
|
return self.loss_accumulator.as_dict()
|