report current loss scale and adjust grad norm by loss scale (for deepspeed)
This commit is contained in:
parent
d0ebce6bac
commit
827cf632e7
|
@ -491,6 +491,12 @@ class Engines(dict[str, Engine]):
|
||||||
|
|
||||||
elapsed_time = time.time() - start_time
|
elapsed_time = time.time() - start_time
|
||||||
total_elapsed_time += elapsed_time
|
total_elapsed_time += elapsed_time
|
||||||
|
grad_norm = engine.get_global_grad_norm()
|
||||||
|
loss_scale = 1
|
||||||
|
if hasattr(engine.optimizer, "loss_scale"):
|
||||||
|
loss_scale = engine.optimizer.loss_scale
|
||||||
|
|
||||||
|
grad_norm /= loss_scale
|
||||||
|
|
||||||
stats.update(
|
stats.update(
|
||||||
flatten_dict(
|
flatten_dict(
|
||||||
|
@ -498,7 +504,8 @@ class Engines(dict[str, Engine]):
|
||||||
name.split("-")[0]: dict(
|
name.split("-")[0]: dict(
|
||||||
**engine_stats,
|
**engine_stats,
|
||||||
lr=engine.get_lr()[0],
|
lr=engine.get_lr()[0],
|
||||||
grad_norm=engine.get_global_grad_norm(),
|
grad_norm=grad_norm,
|
||||||
|
loss_scale=loss_scale if loss_scale != 1 else None,
|
||||||
elapsed_time=elapsed_time,
|
elapsed_time=elapsed_time,
|
||||||
engine_step=engine.global_step,
|
engine_step=engine.global_step,
|
||||||
samples_processed=engine.global_samples,
|
samples_processed=engine.global_samples,
|
||||||
|
|
Loading…
Reference in New Issue
Block a user