Unverified Commit 27b3ff31 authored by Julien Plu's avatar Julien Plu Committed by GitHub
Browse files

Try to understand and apply Sylvain's comments (#8458)

parent 0fa03498
...@@ -23,6 +23,7 @@ from typing import Dict, List, Optional ...@@ -23,6 +23,7 @@ from typing import Dict, List, Optional
import numpy as np import numpy as np
import torch import torch
import transformers
from transformers import ( from transformers import (
AutoConfig, AutoConfig,
AutoModelForSequenceClassification, AutoModelForSequenceClassification,
...@@ -33,6 +34,7 @@ from transformers import ( ...@@ -33,6 +34,7 @@ from transformers import (
default_data_collator, default_data_collator,
set_seed, set_seed,
) )
from transformers.trainer_utils import is_main_process
from utils_hans import HansDataset, InputFeatures, hans_processors, hans_tasks_num_labels from utils_hans import HansDataset, InputFeatures, hans_processors, hans_tasks_num_labels
...@@ -124,6 +126,11 @@ def main(): ...@@ -124,6 +126,11 @@ def main():
bool(training_args.local_rank != -1), bool(training_args.local_rank != -1),
training_args.fp16, training_args.fp16,
) )
# Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(training_args.local_rank):
transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
logger.info("Training/evaluation parameters %s", training_args) logger.info("Training/evaluation parameters %s", training_args)
# Set seed # Set seed
......
...@@ -29,6 +29,7 @@ from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, Tenso ...@@ -29,6 +29,7 @@ from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, Tenso
from torch.utils.data.distributed import DistributedSampler from torch.utils.data.distributed import DistributedSampler
from tqdm import tqdm, trange from tqdm import tqdm, trange
import transformers
from pabee.modeling_pabee_albert import AlbertForSequenceClassificationWithPabee from pabee.modeling_pabee_albert import AlbertForSequenceClassificationWithPabee
from pabee.modeling_pabee_bert import BertForSequenceClassificationWithPabee from pabee.modeling_pabee_bert import BertForSequenceClassificationWithPabee
from transformers import ( from transformers import (
...@@ -44,6 +45,7 @@ from transformers import glue_compute_metrics as compute_metrics ...@@ -44,6 +45,7 @@ from transformers import glue_compute_metrics as compute_metrics
from transformers import glue_convert_examples_to_features as convert_examples_to_features from transformers import glue_convert_examples_to_features as convert_examples_to_features
from transformers import glue_output_modes as output_modes from transformers import glue_output_modes as output_modes
from transformers import glue_processors as processors from transformers import glue_processors as processors
from transformers.trainer_utils import is_main_process
try: try:
...@@ -630,7 +632,11 @@ def main(): ...@@ -630,7 +632,11 @@ def main():
bool(args.local_rank != -1), bool(args.local_rank != -1),
args.fp16, args.fp16,
) )
# Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(args.local_rank):
transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
# Set seed # Set seed
set_seed(args) set_seed(args)
......
...@@ -30,6 +30,7 @@ from torch.utils.data import DataLoader, SequentialSampler, Subset ...@@ -30,6 +30,7 @@ from torch.utils.data import DataLoader, SequentialSampler, Subset
from torch.utils.data.distributed import DistributedSampler from torch.utils.data.distributed import DistributedSampler
from tqdm import tqdm from tqdm import tqdm
import transformers
from transformers import ( from transformers import (
AutoConfig, AutoConfig,
AutoModelForSequenceClassification, AutoModelForSequenceClassification,
...@@ -41,6 +42,7 @@ from transformers import ( ...@@ -41,6 +42,7 @@ from transformers import (
glue_processors, glue_processors,
set_seed, set_seed,
) )
from transformers.trainer_utils import is_main_process
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
...@@ -368,6 +370,11 @@ def main(): ...@@ -368,6 +370,11 @@ def main():
# Setup logging # Setup logging
logging.basicConfig(level=logging.INFO if args.local_rank in [-1, 0] else logging.WARN) logging.basicConfig(level=logging.INFO if args.local_rank in [-1, 0] else logging.WARN)
logger.info("device: {} n_gpu: {}, distributed: {}".format(args.device, args.n_gpu, bool(args.local_rank != -1))) logger.info("device: {} n_gpu: {}, distributed: {}".format(args.device, args.n_gpu, bool(args.local_rank != -1)))
# Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(args.local_rank):
transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
# Set seeds # Set seeds
set_seed(args.seed) set_seed(args.seed)
......
...@@ -29,6 +29,7 @@ from typing import Optional ...@@ -29,6 +29,7 @@ from typing import Optional
from torch.utils.data import ConcatDataset from torch.utils.data import ConcatDataset
import transformers
from transformers import ( from transformers import (
CONFIG_MAPPING, CONFIG_MAPPING,
MODEL_WITH_LM_HEAD_MAPPING, MODEL_WITH_LM_HEAD_MAPPING,
...@@ -47,6 +48,7 @@ from transformers import ( ...@@ -47,6 +48,7 @@ from transformers import (
TrainingArguments, TrainingArguments,
set_seed, set_seed,
) )
from transformers.trainer_utils import is_main_process
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
...@@ -219,6 +221,11 @@ def main(): ...@@ -219,6 +221,11 @@ def main():
bool(training_args.local_rank != -1), bool(training_args.local_rank != -1),
training_args.fp16, training_args.fp16,
) )
# Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(training_args.local_rank):
transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
logger.info("Training/evaluation parameters %s", training_args) logger.info("Training/evaluation parameters %s", training_args)
# Set seed # Set seed
......
...@@ -31,6 +31,7 @@ from torch.utils.data import DataLoader, RandomSampler, SequentialSampler ...@@ -31,6 +31,7 @@ from torch.utils.data import DataLoader, RandomSampler, SequentialSampler
from torch.utils.data.distributed import DistributedSampler from torch.utils.data.distributed import DistributedSampler
from tqdm import tqdm, trange from tqdm import tqdm, trange
import transformers
from transformers import ( from transformers import (
WEIGHTS_NAME, WEIGHTS_NAME,
AdamW, AdamW,
...@@ -41,6 +42,7 @@ from transformers import ( ...@@ -41,6 +42,7 @@ from transformers import (
MMBTForClassification, MMBTForClassification,
get_linear_schedule_with_warmup, get_linear_schedule_with_warmup,
) )
from transformers.trainer_utils import is_main_process
from utils_mmimdb import ImageEncoder, JsonlDataset, collate_fn, get_image_transforms, get_mmimdb_labels from utils_mmimdb import ImageEncoder, JsonlDataset, collate_fn, get_image_transforms, get_mmimdb_labels
...@@ -476,7 +478,11 @@ def main(): ...@@ -476,7 +478,11 @@ def main():
bool(args.local_rank != -1), bool(args.local_rank != -1),
args.fp16, args.fp16,
) )
# Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(args.local_rank):
transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
# Set seed # Set seed
set_seed(args) set_seed(args)
......
...@@ -31,8 +31,10 @@ from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, Tenso ...@@ -31,8 +31,10 @@ from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, Tenso
from torch.utils.data.distributed import DistributedSampler from torch.utils.data.distributed import DistributedSampler
from tqdm import tqdm, trange from tqdm import tqdm, trange
import transformers
from transformers import WEIGHTS_NAME, AdamW, AutoConfig, AutoTokenizer, get_linear_schedule_with_warmup from transformers import WEIGHTS_NAME, AdamW, AutoConfig, AutoTokenizer, get_linear_schedule_with_warmup
from transformers.modeling_auto import AutoModelForMultipleChoice from transformers.modeling_auto import AutoModelForMultipleChoice
from transformers.trainer_utils import is_main_process
try: try:
...@@ -620,6 +622,11 @@ def main(): ...@@ -620,6 +622,11 @@ def main():
bool(args.local_rank != -1), bool(args.local_rank != -1),
args.fp16, args.fp16,
) )
# Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(args.local_rank):
transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
# Set seed # Set seed
set_seed(args) set_seed(args)
......
...@@ -13,6 +13,7 @@ from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, Tenso ...@@ -13,6 +13,7 @@ from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, Tenso
from torch.utils.data.distributed import DistributedSampler from torch.utils.data.distributed import DistributedSampler
from tqdm import tqdm, trange from tqdm import tqdm, trange
import transformers
from src.modeling_highway_bert import DeeBertForSequenceClassification from src.modeling_highway_bert import DeeBertForSequenceClassification
from src.modeling_highway_roberta import DeeRobertaForSequenceClassification from src.modeling_highway_roberta import DeeRobertaForSequenceClassification
from transformers import ( from transformers import (
...@@ -28,6 +29,7 @@ from transformers import glue_compute_metrics as compute_metrics ...@@ -28,6 +29,7 @@ from transformers import glue_compute_metrics as compute_metrics
from transformers import glue_convert_examples_to_features as convert_examples_to_features from transformers import glue_convert_examples_to_features as convert_examples_to_features
from transformers import glue_output_modes as output_modes from transformers import glue_output_modes as output_modes
from transformers import glue_processors as processors from transformers import glue_processors as processors
from transformers.trainer_utils import is_main_process
try: try:
...@@ -580,7 +582,11 @@ def main(): ...@@ -580,7 +582,11 @@ def main():
bool(args.local_rank != -1), bool(args.local_rank != -1),
args.fp16, args.fp16,
) )
# Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(args.local_rank):
transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
# Set seed # Set seed
set_seed(args) set_seed(args)
......
...@@ -30,6 +30,7 @@ from torch.utils.data import DataLoader, RandomSampler, SequentialSampler ...@@ -30,6 +30,7 @@ from torch.utils.data import DataLoader, RandomSampler, SequentialSampler
from torch.utils.data.distributed import DistributedSampler from torch.utils.data.distributed import DistributedSampler
from tqdm import tqdm, trange from tqdm import tqdm, trange
import transformers
from transformers import ( from transformers import (
WEIGHTS_NAME, WEIGHTS_NAME,
AdamW, AdamW,
...@@ -57,6 +58,7 @@ from transformers.data.metrics.squad_metrics import ( ...@@ -57,6 +58,7 @@ from transformers.data.metrics.squad_metrics import (
squad_evaluate, squad_evaluate,
) )
from transformers.data.processors.squad import SquadResult, SquadV1Processor, SquadV2Processor from transformers.data.processors.squad import SquadResult, SquadV1Processor, SquadV2Processor
from transformers.trainer_utils import is_main_process
try: try:
...@@ -745,7 +747,11 @@ def main(): ...@@ -745,7 +747,11 @@ def main():
bool(args.local_rank != -1), bool(args.local_rank != -1),
args.fp16, args.fp16,
) )
# Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(args.local_rank):
transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
# Set seed # Set seed
set_seed(args) set_seed(args)
......
...@@ -168,6 +168,8 @@ def main(): ...@@ -168,6 +168,8 @@ def main():
# Set the verbosity to info of the Transformers logger (on main process only): # Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(training_args.local_rank): if is_main_process(training_args.local_rank):
transformers.utils.logging.set_verbosity_info() transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
logger.info("Training/evaluation parameters %s", training_args) logger.info("Training/evaluation parameters %s", training_args)
# Set seed before initializing model. # Set seed before initializing model.
......
...@@ -179,6 +179,8 @@ def main(): ...@@ -179,6 +179,8 @@ def main():
# Set the verbosity to info of the Transformers logger (on main process only): # Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(training_args.local_rank): if is_main_process(training_args.local_rank):
transformers.utils.logging.set_verbosity_info() transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
logger.info("Training/evaluation parameters %s", training_args) logger.info("Training/evaluation parameters %s", training_args)
# Set seed before initializing model. # Set seed before initializing model.
......
...@@ -186,6 +186,8 @@ def main(): ...@@ -186,6 +186,8 @@ def main():
# Set the verbosity to info of the Transformers logger (on main process only): # Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(training_args.local_rank): if is_main_process(training_args.local_rank):
transformers.utils.logging.set_verbosity_info() transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
logger.info("Training/evaluation parameters %s", training_args) logger.info("Training/evaluation parameters %s", training_args)
# Set seed before initializing model. # Set seed before initializing model.
......
...@@ -176,6 +176,8 @@ def main(): ...@@ -176,6 +176,8 @@ def main():
# Set the verbosity to info of the Transformers logger (on main process only): # Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(training_args.local_rank): if is_main_process(training_args.local_rank):
transformers.utils.logging.set_verbosity_info() transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
logger.info("Training/evaluation parameters %s", training_args) logger.info("Training/evaluation parameters %s", training_args)
# Set seed before initializing model. # Set seed before initializing model.
......
...@@ -23,6 +23,7 @@ from typing import Dict, Optional ...@@ -23,6 +23,7 @@ from typing import Dict, Optional
import numpy as np import numpy as np
import transformers
from transformers import ( from transformers import (
AutoConfig, AutoConfig,
AutoModelForMultipleChoice, AutoModelForMultipleChoice,
...@@ -33,6 +34,7 @@ from transformers import ( ...@@ -33,6 +34,7 @@ from transformers import (
TrainingArguments, TrainingArguments,
set_seed, set_seed,
) )
from transformers.trainer_utils import is_main_process
from utils_multiple_choice import MultipleChoiceDataset, Split, processors from utils_multiple_choice import MultipleChoiceDataset, Split, processors
...@@ -115,6 +117,11 @@ def main(): ...@@ -115,6 +117,11 @@ def main():
bool(training_args.local_rank != -1), bool(training_args.local_rank != -1),
training_args.fp16, training_args.fp16,
) )
# Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(training_args.local_rank):
transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
logger.info("Training/evaluation parameters %s", training_args) logger.info("Training/evaluation parameters %s", training_args)
# Set seed # Set seed
......
...@@ -33,9 +33,15 @@ from transformers import ( ...@@ -33,9 +33,15 @@ from transformers import (
TFTrainingArguments, TFTrainingArguments,
set_seed, set_seed,
) )
from transformers.utils import logging as hf_logging
from utils_multiple_choice import Split, TFMultipleChoiceDataset, processors from utils_multiple_choice import Split, TFMultipleChoiceDataset, processors
hf_logging.set_verbosity_info()
hf_logging.enable_default_handler()
hf_logging.enable_explicit_format()
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
......
...@@ -29,6 +29,7 @@ from torch.utils.data import DataLoader, RandomSampler, SequentialSampler ...@@ -29,6 +29,7 @@ from torch.utils.data import DataLoader, RandomSampler, SequentialSampler
from torch.utils.data.distributed import DistributedSampler from torch.utils.data.distributed import DistributedSampler
from tqdm import tqdm, trange from tqdm import tqdm, trange
import transformers
from transformers import ( from transformers import (
MODEL_FOR_QUESTION_ANSWERING_MAPPING, MODEL_FOR_QUESTION_ANSWERING_MAPPING,
WEIGHTS_NAME, WEIGHTS_NAME,
...@@ -45,6 +46,7 @@ from transformers.data.metrics.squad_metrics import ( ...@@ -45,6 +46,7 @@ from transformers.data.metrics.squad_metrics import (
squad_evaluate, squad_evaluate,
) )
from transformers.data.processors.squad import SquadResult, SquadV1Processor, SquadV2Processor from transformers.data.processors.squad import SquadResult, SquadV1Processor, SquadV2Processor
from transformers.trainer_utils import is_main_process
try: try:
...@@ -712,7 +714,11 @@ def main(): ...@@ -712,7 +714,11 @@ def main():
bool(args.local_rank != -1), bool(args.local_rank != -1),
args.fp16, args.fp16,
) )
# Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(args.local_rank):
transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
# Set seed # Set seed
set_seed(args) set_seed(args)
......
...@@ -22,9 +22,11 @@ import sys ...@@ -22,9 +22,11 @@ import sys
from dataclasses import dataclass, field from dataclasses import dataclass, field
from typing import Optional from typing import Optional
import transformers
from transformers import AutoConfig, AutoModelForQuestionAnswering, AutoTokenizer, HfArgumentParser, SquadDataset from transformers import AutoConfig, AutoModelForQuestionAnswering, AutoTokenizer, HfArgumentParser, SquadDataset
from transformers import SquadDataTrainingArguments as DataTrainingArguments from transformers import SquadDataTrainingArguments as DataTrainingArguments
from transformers import Trainer, TrainingArguments from transformers import Trainer, TrainingArguments
from transformers.trainer_utils import is_main_process
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
...@@ -91,6 +93,11 @@ def main(): ...@@ -91,6 +93,11 @@ def main():
bool(training_args.local_rank != -1), bool(training_args.local_rank != -1),
training_args.fp16, training_args.fp16,
) )
# Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(training_args.local_rank):
transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
logger.info("Training/evaluation parameters %s", training_args) logger.info("Training/evaluation parameters %s", training_args)
# Prepare Question-Answering task # Prepare Question-Answering task
......
...@@ -33,6 +33,12 @@ from transformers import ( ...@@ -33,6 +33,12 @@ from transformers import (
squad_convert_examples_to_features, squad_convert_examples_to_features,
) )
from transformers.data.processors.squad import SquadV1Processor, SquadV2Processor from transformers.data.processors.squad import SquadV1Processor, SquadV2Processor
from transformers.utils import logging as hf_logging
hf_logging.set_verbosity_info()
hf_logging.enable_default_handler()
hf_logging.enable_explicit_format()
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
......
...@@ -4,10 +4,11 @@ import sys ...@@ -4,10 +4,11 @@ import sys
from dataclasses import dataclass, field from dataclasses import dataclass, field
from typing import Optional from typing import Optional
import transformers
from seq2seq_trainer import Seq2SeqTrainer from seq2seq_trainer import Seq2SeqTrainer
from seq2seq_training_args import Seq2SeqTrainingArguments from seq2seq_training_args import Seq2SeqTrainingArguments
from transformers import AutoConfig, AutoModelForSeq2SeqLM, AutoTokenizer, HfArgumentParser, MBartTokenizer, set_seed from transformers import AutoConfig, AutoModelForSeq2SeqLM, AutoTokenizer, HfArgumentParser, MBartTokenizer, set_seed
from transformers.trainer_utils import EvaluationStrategy from transformers.trainer_utils import EvaluationStrategy, is_main_process
from utils import ( from utils import (
Seq2SeqDataCollator, Seq2SeqDataCollator,
Seq2SeqDataset, Seq2SeqDataset,
...@@ -131,6 +132,11 @@ def main(): ...@@ -131,6 +132,11 @@ def main():
bool(training_args.local_rank != -1), bool(training_args.local_rank != -1),
training_args.fp16, training_args.fp16,
) )
# Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(training_args.local_rank):
transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
logger.info("Training/evaluation parameters %s", training_args) logger.info("Training/evaluation parameters %s", training_args)
# Set seed # Set seed
......
...@@ -171,6 +171,8 @@ def main(): ...@@ -171,6 +171,8 @@ def main():
# Set the verbosity to info of the Transformers logger (on main process only): # Set the verbosity to info of the Transformers logger (on main process only):
if is_main_process(training_args.local_rank): if is_main_process(training_args.local_rank):
transformers.utils.logging.set_verbosity_info() transformers.utils.logging.set_verbosity_info()
transformers.utils.logging.enable_default_handler()
transformers.utils.logging.enable_explicit_format()
logger.info(f"Training/evaluation parameters {training_args}") logger.info(f"Training/evaluation parameters {training_args}")
# Set seed before initializing model. # Set seed before initializing model.
......
...@@ -27,6 +27,12 @@ from transformers import ( ...@@ -27,6 +27,12 @@ from transformers import (
glue_processors, glue_processors,
glue_tasks_num_labels, glue_tasks_num_labels,
) )
from transformers.utils import logging as hf_logging
hf_logging.set_verbosity_info()
hf_logging.enable_default_handler()
hf_logging.enable_explicit_format()
class Split(Enum): class Split(Enum):
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment