"...git@developer.sourcefind.cn:cnjsdfcy/simbricks.git" did not exist on "07757fdc2de320f44fa5465e821d468a1951579c"
Unverified Commit deb2b590 authored by bofeng huang's avatar bofeng huang Committed by GitHub
Browse files

Fix lr_scheduler in no_trainer training scripts (#27872)

* Fix lr_scheduler

* Fix lr scheduler
parent 692c3c6b
...@@ -438,8 +438,8 @@ def main(): ...@@ -438,8 +438,8 @@ def main():
lr_scheduler = get_scheduler( lr_scheduler = get_scheduler(
name=args.lr_scheduler_type, name=args.lr_scheduler_type,
optimizer=optimizer, optimizer=optimizer,
num_warmup_steps=args.num_warmup_steps * args.gradient_accumulation_steps, num_warmup_steps=args.num_warmup_steps * accelerator.num_processes,
num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, num_training_steps=args.max_train_steps if overrode_max_train_steps else args.max_train_steps * accelerator.num_processes,
) )
# Prepare everything with our `accelerator`. # Prepare everything with our `accelerator`.
......
...@@ -626,8 +626,8 @@ def main(): ...@@ -626,8 +626,8 @@ def main():
lr_scheduler = get_scheduler( lr_scheduler = get_scheduler(
name=args.lr_scheduler_type, name=args.lr_scheduler_type,
optimizer=optimizer, optimizer=optimizer,
num_warmup_steps=args.num_warmup_steps * args.gradient_accumulation_steps, num_warmup_steps=args.num_warmup_steps * accelerator.num_processes,
num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, num_training_steps=args.max_train_steps if overrode_max_train_steps else args.max_train_steps * accelerator.num_processes,
) )
# Prepare everything with our `accelerator`. # Prepare everything with our `accelerator`.
......
...@@ -526,8 +526,8 @@ def main(): ...@@ -526,8 +526,8 @@ def main():
lr_scheduler = get_scheduler( lr_scheduler = get_scheduler(
name=args.lr_scheduler_type, name=args.lr_scheduler_type,
optimizer=optimizer, optimizer=optimizer,
num_warmup_steps=args.num_warmup_steps * args.gradient_accumulation_steps, num_warmup_steps=args.num_warmup_steps * accelerator.num_processes,
num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, num_training_steps=args.max_train_steps if overrode_max_train_steps else args.max_train_steps * accelerator.num_processes,
) )
# Prepare everything with our `accelerator`. # Prepare everything with our `accelerator`.
......
...@@ -563,8 +563,8 @@ def main(): ...@@ -563,8 +563,8 @@ def main():
lr_scheduler = get_scheduler( lr_scheduler = get_scheduler(
name=args.lr_scheduler_type, name=args.lr_scheduler_type,
optimizer=optimizer, optimizer=optimizer,
num_warmup_steps=args.num_warmup_steps * args.gradient_accumulation_steps, num_warmup_steps=args.num_warmup_steps * accelerator.num_processes,
num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, num_training_steps=args.max_train_steps if overrode_max_train_steps else args.max_train_steps * accelerator.num_processes,
) )
# Prepare everything with our `accelerator`. # Prepare everything with our `accelerator`.
......
...@@ -510,8 +510,8 @@ def main(): ...@@ -510,8 +510,8 @@ def main():
lr_scheduler = get_scheduler( lr_scheduler = get_scheduler(
name=args.lr_scheduler_type, name=args.lr_scheduler_type,
optimizer=optimizer, optimizer=optimizer,
num_warmup_steps=args.num_warmup_steps * args.gradient_accumulation_steps, num_warmup_steps=args.num_warmup_steps * accelerator.num_processes,
num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, num_training_steps=args.max_train_steps if overrode_max_train_steps else args.max_train_steps * accelerator.num_processes,
) )
# Prepare everything with our `accelerator`. # Prepare everything with our `accelerator`.
......
...@@ -750,8 +750,8 @@ def main(): ...@@ -750,8 +750,8 @@ def main():
lr_scheduler = get_scheduler( lr_scheduler = get_scheduler(
name=args.lr_scheduler_type, name=args.lr_scheduler_type,
optimizer=optimizer, optimizer=optimizer,
num_warmup_steps=args.num_warmup_steps * args.gradient_accumulation_steps, num_warmup_steps=args.num_warmup_steps * accelerator.num_processes,
num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, num_training_steps=args.max_train_steps if overrode_max_train_steps else args.max_train_steps * accelerator.num_processes,
) )
# Prepare everything with our `accelerator`. # Prepare everything with our `accelerator`.
......
...@@ -780,8 +780,8 @@ def main(): ...@@ -780,8 +780,8 @@ def main():
lr_scheduler = get_scheduler( lr_scheduler = get_scheduler(
name=args.lr_scheduler_type, name=args.lr_scheduler_type,
optimizer=optimizer, optimizer=optimizer,
num_warmup_steps=args.num_warmup_steps * args.gradient_accumulation_steps, num_warmup_steps=args.num_warmup_steps * accelerator.num_processes,
num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, num_training_steps=args.max_train_steps if overrode_max_train_steps else args.max_train_steps * accelerator.num_processes,
) )
# Prepare everything with our `accelerator`. # Prepare everything with our `accelerator`.
......
...@@ -513,8 +513,8 @@ def main(): ...@@ -513,8 +513,8 @@ def main():
lr_scheduler = get_scheduler( lr_scheduler = get_scheduler(
name=args.lr_scheduler_type, name=args.lr_scheduler_type,
optimizer=optimizer, optimizer=optimizer,
num_warmup_steps=args.num_warmup_steps * args.gradient_accumulation_steps, num_warmup_steps=args.num_warmup_steps * accelerator.num_processes,
num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, num_training_steps=args.max_train_steps if overrode_max_train_steps else args.max_train_steps * accelerator.num_processes,
) )
# Prepare everything with our `accelerator`. # Prepare everything with our `accelerator`.
......
...@@ -580,8 +580,8 @@ def main(): ...@@ -580,8 +580,8 @@ def main():
lr_scheduler = get_scheduler( lr_scheduler = get_scheduler(
name=args.lr_scheduler_type, name=args.lr_scheduler_type,
optimizer=optimizer, optimizer=optimizer,
num_warmup_steps=args.num_warmup_steps * args.gradient_accumulation_steps, num_warmup_steps=args.num_warmup_steps * accelerator.num_processes,
num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, num_training_steps=args.max_train_steps if overrode_max_train_steps else args.max_train_steps * accelerator.num_processes,
) )
# Prepare everything with our `accelerator`. # Prepare everything with our `accelerator`.
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment