From e7957e489764f4fb1cf86d869e7120ff542d45fe Mon Sep 17 00:00:00 2001 From: James Betker Date: Sun, 12 Dec 2021 22:23:17 -0700 Subject: [PATCH] Make loss accumulator for logs accumulate better --- codes/trainer/steps.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/codes/trainer/steps.py b/codes/trainer/steps.py index 20ac0a92..13ac185e 100644 --- a/codes/trainer/steps.py +++ b/codes/trainer/steps.py @@ -22,7 +22,7 @@ class ConfigurableStep(Module): self.env = env self.opt = env['opt'] self.gen_outputs = opt_step['generator_outputs'] - self.loss_accumulator = LossAccumulator() + self.loss_accumulator = LossAccumulator(buffer_sz=opt_get(opt_step, ['loss_log_buffer'], 50)) self.optimizers = None self.scaler = GradScaler(enabled=self.opt['fp16']) self.grads_generated = False