Unverified Commit f270b960 authored by Joao Gante's avatar Joao Gante Committed by GitHub
Browse files

Generate: move generation_*.py src files into generation/*.py (#20096)

* move generation_*.py src files into generation/*.py

* populate generation.__init__ with lazy loading

* move imports and references from generation.xxx.object to generation.object
parent bac2d29a
...@@ -31,6 +31,13 @@ class TFForcedEOSTokenLogitsProcessor(metaclass=DummyObject): ...@@ -31,6 +31,13 @@ class TFForcedEOSTokenLogitsProcessor(metaclass=DummyObject):
requires_backends(self, ["tf"]) requires_backends(self, ["tf"])
class TFGenerationMixin(metaclass=DummyObject):
_backends = ["tf"]
def __init__(self, *args, **kwargs):
requires_backends(self, ["tf"])
class TFLogitsProcessor(metaclass=DummyObject): class TFLogitsProcessor(metaclass=DummyObject):
_backends = ["tf"] _backends = ["tf"]
......
...@@ -490,7 +490,7 @@ from transformers.utils import cached_property ...@@ -490,7 +490,7 @@ from transformers.utils import cached_property
from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device
from ...test_configuration_common import ConfigTester from ...test_configuration_common import ConfigTester
from ...generation.test_generation_utils import GenerationTesterMixin from ...generation.test_utils import GenerationTesterMixin
from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_modeling_common import ModelTesterMixin, ids_tensor
......
...@@ -23,7 +23,7 @@ from transformers.testing_utils import require_torch ...@@ -23,7 +23,7 @@ from transformers.testing_utils import require_torch
if is_torch_available(): if is_torch_available():
import torch import torch
from transformers.generation_beam_constraints import DisjunctiveConstraint from transformers.generation import DisjunctiveConstraint
@require_torch @require_torch
......
...@@ -25,8 +25,13 @@ from ..test_modeling_common import floats_tensor, ids_tensor ...@@ -25,8 +25,13 @@ from ..test_modeling_common import floats_tensor, ids_tensor
if is_torch_available(): if is_torch_available():
import torch import torch
from transformers.generation_beam_constraints import DisjunctiveConstraint, PhrasalConstraint from transformers.generation import (
from transformers.generation_beam_search import BeamHypotheses, BeamSearchScorer, ConstrainedBeamSearchScorer BeamHypotheses,
BeamSearchScorer,
ConstrainedBeamSearchScorer,
DisjunctiveConstraint,
PhrasalConstraint,
)
class BeamSearchTester: class BeamSearchTester:
......
...@@ -27,7 +27,7 @@ from ..test_modeling_flax_common import ids_tensor ...@@ -27,7 +27,7 @@ from ..test_modeling_flax_common import ids_tensor
if is_flax_available(): if is_flax_available():
import jax import jax
import jax.numpy as jnp import jax.numpy as jnp
from transformers.generation_flax_logits_process import ( from transformers.generation import (
FlaxForcedBOSTokenLogitsProcessor, FlaxForcedBOSTokenLogitsProcessor,
FlaxForcedEOSTokenLogitsProcessor, FlaxForcedEOSTokenLogitsProcessor,
FlaxLogitsProcessorList, FlaxLogitsProcessorList,
......
...@@ -26,7 +26,7 @@ if is_torch_available(): ...@@ -26,7 +26,7 @@ if is_torch_available():
import torch import torch
from torch import nn from torch import nn
from transformers.generation_logits_process import ( from transformers.generation import (
EncoderNoRepeatNGramLogitsProcessor, EncoderNoRepeatNGramLogitsProcessor,
ExponentialDecayLengthPenalty, ExponentialDecayLengthPenalty,
ForcedBOSTokenLogitsProcessor, ForcedBOSTokenLogitsProcessor,
......
...@@ -25,7 +25,7 @@ from ..test_modeling_common import ids_tensor ...@@ -25,7 +25,7 @@ from ..test_modeling_common import ids_tensor
if is_torch_available(): if is_torch_available():
import torch import torch
from transformers.generation_stopping_criteria import ( from transformers.generation import (
MaxLengthCriteria, MaxLengthCriteria,
MaxNewTokensCriteria, MaxNewTokensCriteria,
MaxTimeCriteria, MaxTimeCriteria,
......
...@@ -26,7 +26,7 @@ from transformers.testing_utils import require_tf ...@@ -26,7 +26,7 @@ from transformers.testing_utils import require_tf
if is_tf_available(): if is_tf_available():
import tensorflow as tf import tensorflow as tf
from transformers.generation_tf_logits_process import ( from transformers.generation import (
TFForcedBOSTokenLogitsProcessor, TFForcedBOSTokenLogitsProcessor,
TFForcedEOSTokenLogitsProcessor, TFForcedEOSTokenLogitsProcessor,
TFForceTokensLogitsProcessor, TFForceTokensLogitsProcessor,
......
...@@ -42,33 +42,35 @@ if is_torch_available(): ...@@ -42,33 +42,35 @@ if is_torch_available():
pipeline, pipeline,
top_k_top_p_filtering, top_k_top_p_filtering,
) )
from transformers.generation_beam_constraints import DisjunctiveConstraint, PhrasalConstraint from transformers.generation import (
from transformers.generation_beam_search import BeamSearchScorer, ConstrainedBeamSearchScorer BeamSampleDecoderOnlyOutput,
from transformers.generation_logits_process import ( BeamSampleEncoderDecoderOutput,
BeamSearchDecoderOnlyOutput,
BeamSearchEncoderDecoderOutput,
BeamSearchScorer,
ConstrainedBeamSearchScorer,
DisjunctiveConstraint,
ForcedBOSTokenLogitsProcessor, ForcedBOSTokenLogitsProcessor,
ForcedEOSTokenLogitsProcessor, ForcedEOSTokenLogitsProcessor,
GreedySearchDecoderOnlyOutput,
GreedySearchEncoderDecoderOutput,
HammingDiversityLogitsProcessor, HammingDiversityLogitsProcessor,
InfNanRemoveLogitsProcessor, InfNanRemoveLogitsProcessor,
LogitsProcessorList, LogitsProcessorList,
MaxLengthCriteria,
MinLengthLogitsProcessor, MinLengthLogitsProcessor,
NoBadWordsLogitsProcessor, NoBadWordsLogitsProcessor,
NoRepeatNGramLogitsProcessor, NoRepeatNGramLogitsProcessor,
PhrasalConstraint,
RepetitionPenaltyLogitsProcessor, RepetitionPenaltyLogitsProcessor,
SampleDecoderOnlyOutput,
SampleEncoderDecoderOutput,
StoppingCriteria,
StoppingCriteriaList,
TemperatureLogitsWarper, TemperatureLogitsWarper,
TopKLogitsWarper, TopKLogitsWarper,
TopPLogitsWarper, TopPLogitsWarper,
) )
from transformers.generation_stopping_criteria import MaxLengthCriteria, StoppingCriteria, StoppingCriteriaList
from transformers.generation_utils import (
BeamSampleDecoderOnlyOutput,
BeamSampleEncoderDecoderOutput,
BeamSearchDecoderOnlyOutput,
BeamSearchEncoderDecoderOutput,
GreedySearchDecoderOnlyOutput,
GreedySearchEncoderDecoderOutput,
SampleDecoderOnlyOutput,
SampleEncoderDecoderOutput,
)
class GenerationTesterMixin: class GenerationTesterMixin:
......
...@@ -25,7 +25,7 @@ from transformers import BartConfig, is_torch_available ...@@ -25,7 +25,7 @@ from transformers import BartConfig, is_torch_available
from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device
from transformers.utils import cached_property from transformers.utils import cached_property
from ...generation.test_generation_utils import GenerationTesterMixin from ...generation.test_utils import GenerationTesterMixin
from ...test_configuration_common import ConfigTester from ...test_configuration_common import ConfigTester
from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor
......
...@@ -19,7 +19,7 @@ import timeout_decorator # noqa ...@@ -19,7 +19,7 @@ import timeout_decorator # noqa
from transformers import BartConfig, BartTokenizer, is_flax_available from transformers import BartConfig, BartTokenizer, is_flax_available
from transformers.testing_utils import require_flax, slow from transformers.testing_utils import require_flax, slow
from ...generation.test_generation_flax_utils import FlaxGenerationTesterMixin from ...generation.test_flax_utils import FlaxGenerationTesterMixin
from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask
......
...@@ -20,7 +20,7 @@ from transformers import BertConfig, is_torch_available ...@@ -20,7 +20,7 @@ from transformers import BertConfig, is_torch_available
from transformers.models.auto import get_values from transformers.models.auto import get_values
from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device
from ...generation.test_generation_utils import GenerationTesterMixin from ...generation.test_utils import GenerationTesterMixin
from ...test_configuration_common import ConfigTester from ...test_configuration_common import ConfigTester
from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask
......
...@@ -19,7 +19,7 @@ import unittest ...@@ -19,7 +19,7 @@ import unittest
from transformers import BertGenerationConfig, is_torch_available from transformers import BertGenerationConfig, is_torch_available
from transformers.testing_utils import require_torch, slow, torch_device from transformers.testing_utils import require_torch, slow, torch_device
from ...generation.test_generation_utils import GenerationTesterMixin from ...generation.test_utils import GenerationTesterMixin
from ...test_configuration_common import ConfigTester from ...test_configuration_common import ConfigTester
from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask
......
...@@ -22,7 +22,7 @@ import unittest ...@@ -22,7 +22,7 @@ import unittest
from transformers import BigBirdPegasusConfig, is_torch_available from transformers import BigBirdPegasusConfig, is_torch_available
from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device
from ...generation.test_generation_utils import GenerationTesterMixin from ...generation.test_utils import GenerationTesterMixin
from ...test_configuration_common import ConfigTester from ...test_configuration_common import ConfigTester
from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_modeling_common import ModelTesterMixin, ids_tensor
......
...@@ -21,7 +21,7 @@ from transformers import BlenderbotConfig, is_torch_available ...@@ -21,7 +21,7 @@ from transformers import BlenderbotConfig, is_torch_available
from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device
from transformers.utils import cached_property from transformers.utils import cached_property
from ...generation.test_generation_utils import GenerationTesterMixin from ...generation.test_utils import GenerationTesterMixin
from ...test_configuration_common import ConfigTester from ...test_configuration_common import ConfigTester
from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_modeling_common import ModelTesterMixin, ids_tensor
......
...@@ -20,7 +20,7 @@ import timeout_decorator # noqa ...@@ -20,7 +20,7 @@ import timeout_decorator # noqa
from transformers import BlenderbotConfig, is_flax_available from transformers import BlenderbotConfig, is_flax_available
from transformers.testing_utils import jax_device, require_flax, slow from transformers.testing_utils import jax_device, require_flax, slow
from ...generation.test_generation_flax_utils import FlaxGenerationTesterMixin from ...generation.test_flax_utils import FlaxGenerationTesterMixin
from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor
......
...@@ -21,7 +21,7 @@ from transformers import BlenderbotSmallConfig, is_torch_available ...@@ -21,7 +21,7 @@ from transformers import BlenderbotSmallConfig, is_torch_available
from transformers.testing_utils import require_torch, slow, torch_device from transformers.testing_utils import require_torch, slow, torch_device
from transformers.utils import cached_property from transformers.utils import cached_property
from ...generation.test_generation_utils import GenerationTesterMixin from ...generation.test_utils import GenerationTesterMixin
from ...test_configuration_common import ConfigTester from ...test_configuration_common import ConfigTester
from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_modeling_common import ModelTesterMixin, ids_tensor
......
...@@ -20,7 +20,7 @@ import timeout_decorator # noqa ...@@ -20,7 +20,7 @@ import timeout_decorator # noqa
from transformers import BlenderbotSmallConfig, is_flax_available from transformers import BlenderbotSmallConfig, is_flax_available
from transformers.testing_utils import require_flax, slow from transformers.testing_utils import require_flax, slow
from ...generation.test_generation_flax_utils import FlaxGenerationTesterMixin from ...generation.test_flax_utils import FlaxGenerationTesterMixin
from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment