Unverified Commit 67cd4601 authored by Will Berman's avatar Will Berman Committed by GitHub
Browse files

do not scale the initial global step by gradient accumulation steps when...

do not scale the initial global step by gradient accumulation steps when loading from checkpoint (#3506)
parent 64bf5d33
......@@ -979,7 +979,7 @@ def main(args):
accelerator.load_state(os.path.join(args.output_dir, path))
global_step = int(path.split("-")[1])
initial_global_step = global_step * args.gradient_accumulation_steps
initial_global_step = global_step
first_epoch = global_step // num_update_steps_per_epoch
else:
initial_global_step = 0
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment