From 9cf6c3824bf0fa21dbe1ec375dfe41a6a1b21ae1 Mon Sep 17 00:00:00 2001 From: Peter-1213 Date: Wed, 27 Oct 2021 01:22:33 +0800 Subject: [PATCH] Fix a logic flaw that causes checkpoints to be saved at wrong epochs. --- PyTorch/Translation/Transformer/train.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/PyTorch/Translation/Transformer/train.py b/PyTorch/Translation/Transformer/train.py index 98147d8a..a749e535 100644 --- a/PyTorch/Translation/Transformer/train.py +++ b/PyTorch/Translation/Transformer/train.py @@ -364,7 +364,7 @@ def _all_gather_predictions(predictions): def save_checkpoint(args, trainer, epoch_itr, val_loss): - if epoch_itr.epoch % args.save_interval == 0: + if epoch_itr.epoch % args.save_interval != 0: return if args.no_save or not distributed_utils.is_master(args): return