"...git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "691fd8fdded34497451bc638716eb8fbd484883a"
Unverified Commit e2770748 authored by Stas Bekman's avatar Stas Bekman Committed by GitHub
Browse files

pass the matching trainer log level to deepspeed (#12401)

parent 7e22609e
...@@ -295,11 +295,13 @@ def deepspeed_init(trainer, num_training_steps, resume_from_checkpoint=None): ...@@ -295,11 +295,13 @@ def deepspeed_init(trainer, num_training_steps, resume_from_checkpoint=None):
""" """
import deepspeed import deepspeed
from deepspeed.utils import logger as ds_logger
model = trainer.model model = trainer.model
args = trainer.args
hf_deepspeed_config = trainer.args.hf_deepspeed_config hf_deepspeed_config = args.hf_deepspeed_config
hf_deepspeed_config.trainer_config_finalize(trainer.args, model, num_training_steps) hf_deepspeed_config.trainer_config_finalize(args, model, num_training_steps)
# resume config update - some bits like `model` and `num_training_steps` only become available during train # resume config update - some bits like `model` and `num_training_steps` only become available during train
config = hf_deepspeed_config.config config = hf_deepspeed_config.config
...@@ -319,7 +321,7 @@ def deepspeed_init(trainer, num_training_steps, resume_from_checkpoint=None): ...@@ -319,7 +321,7 @@ def deepspeed_init(trainer, num_training_steps, resume_from_checkpoint=None):
optimizer = None optimizer = None
if "optimizer" in config: if "optimizer" in config:
if trainer.args.adafactor: if args.adafactor:
raise ValueError( raise ValueError(
"--adafactor was passed, but also found `optimizer` configured in the DeepSpeed config. " "--adafactor was passed, but also found `optimizer` configured in the DeepSpeed config. "
"Only one optimizer can be configured." "Only one optimizer can be configured."
...@@ -356,6 +358,9 @@ def deepspeed_init(trainer, num_training_steps, resume_from_checkpoint=None): ...@@ -356,6 +358,9 @@ def deepspeed_init(trainer, num_training_steps, resume_from_checkpoint=None):
# keep for quick debug: # keep for quick debug:
# from pprint import pprint; pprint(config) # from pprint import pprint; pprint(config)
# set the Deepspeed log level consistent with the trainer
ds_logger.setLevel(args.get_process_log_level())
model_parameters = filter(lambda p: p.requires_grad, model.parameters()) model_parameters = filter(lambda p: p.requires_grad, model.parameters())
model, optimizer, _, lr_scheduler = deepspeed.initialize( model, optimizer, _, lr_scheduler = deepspeed.initialize(
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment