"git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "4603fe9b1f798d81cf55de65a46691b3276edccd"
Unverified Commit 0bf34b1c authored by Yih-Dar's avatar Yih-Dar Committed by GitHub
Browse files

Skip pt/flax equivalence tests in pytorch `bigbird` test file (#23040)



skip
Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
parent 4d0ea3d2
...@@ -20,7 +20,7 @@ import unittest ...@@ -20,7 +20,7 @@ import unittest
from transformers import BigBirdConfig, is_torch_available from transformers import BigBirdConfig, is_torch_available
from transformers.models.auto import get_values from transformers.models.auto import get_values
from transformers.models.big_bird.tokenization_big_bird import BigBirdTokenizer from transformers.models.big_bird.tokenization_big_bird import BigBirdTokenizer
from transformers.testing_utils import require_torch, slow, torch_device from transformers.testing_utils import is_pt_flax_cross_test, require_torch, slow, torch_device
from ...test_configuration_common import ConfigTester from ...test_configuration_common import ConfigTester
from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask
...@@ -618,6 +618,20 @@ class BigBirdModelTest(ModelTesterMixin, PipelineTesterMixin, unittest.TestCase) ...@@ -618,6 +618,20 @@ class BigBirdModelTest(ModelTesterMixin, PipelineTesterMixin, unittest.TestCase)
else: else:
super().check_pt_flax_outputs(fx_outputs, pt_outputs, model_class, tol, name, attributes) super().check_pt_flax_outputs(fx_outputs, pt_outputs, model_class, tol, name, attributes)
@is_pt_flax_cross_test
@unittest.skip(
reason="Current Pytorch implementation has bug with random attention -> it always uses it not matter if we are in eval/train mode"
)
def test_equivalence_flax_to_pt(self):
pass
@is_pt_flax_cross_test
@unittest.skip(
reason="Current Pytorch implementation has bug with random attention -> it always uses it not matter if we are in eval/train mode"
)
def test_equivalence_pt_to_flax(self):
pass
@require_torch @require_torch
@slow @slow
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment