"tests/models/codegen/test_modeling_codegen.py" did not exist on "41c559415ae72fc84afdf4150981e2db025b9c61"
Unverified Commit 58fb3c9f authored by Yih-Dar's avatar Yih-Dar Committed by GitHub
Browse files

Fix Tapas tests (#17510)


Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
parent ca1f1c86
......@@ -32,7 +32,13 @@ from transformers import (
is_torch_available,
)
from transformers.models.auto import get_values
from transformers.testing_utils import require_scatter, require_torch, slow, torch_device
from transformers.testing_utils import (
require_scatter,
require_tensorflow_probability,
require_torch,
slow,
torch_device,
)
from transformers.utils import cached_property
from ...test_configuration_common import ConfigTester
......@@ -499,6 +505,10 @@ class TapasModelTest(ModelTesterMixin, unittest.TestCase):
config_and_inputs = self.model_tester.prepare_config_and_inputs()
self.model_tester.create_and_check_for_sequence_classification(*config_and_inputs)
@require_tensorflow_probability
def test_pt_tf_model_equivalence(self):
super().test_pt_tf_model_equivalence()
def prepare_tapas_single_inputs_for_inference():
# Here we prepare a single table-question pair to test TAPAS inference on:
......
......@@ -36,6 +36,7 @@ from transformers.testing_utils import (
is_pt_tf_cross_test,
require_pandas,
require_scatter,
require_tensorflow_probability,
require_tokenizers,
require_torch,
slow,
......@@ -141,6 +142,10 @@ class TapasTokenizationTest(TokenizerTesterMixin, unittest.TestCase):
output_text = "unwanted, running"
return input_text, output_text
@require_tensorflow_probability
def test_tf_encode_plus_sent_to_model(self):
super().test_tf_encode_plus_sent_to_model()
def test_rust_and_python_full_tokenizers(self):
if not self.test_rust_tokenizer:
return
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment