Unverified Commit deb72cb6 authored by Yoach Lacombe's avatar Yoach Lacombe Committed by GitHub
Browse files

Skip M4T `test_retain_grad_hidden_states_attentions` (#28060)

* skip test from SpeechInput

* refine description of skip
parent d269c4b2
...@@ -20,7 +20,7 @@ import tempfile ...@@ -20,7 +20,7 @@ import tempfile
import unittest import unittest
from transformers import SeamlessM4TConfig, is_speech_available, is_torch_available from transformers import SeamlessM4TConfig, is_speech_available, is_torch_available
from transformers.testing_utils import is_flaky, require_torch, slow, torch_device from transformers.testing_utils import require_torch, slow, torch_device
from transformers.trainer_utils import set_seed from transformers.trainer_utils import set_seed
from transformers.utils import cached_property from transformers.utils import cached_property
...@@ -610,9 +610,11 @@ class SeamlessM4TModelWithSpeechInputTest(ModelTesterMixin, unittest.TestCase): ...@@ -610,9 +610,11 @@ class SeamlessM4TModelWithSpeechInputTest(ModelTesterMixin, unittest.TestCase):
[self.model_tester.num_attention_heads, encoder_seq_length, encoder_key_length], [self.model_tester.num_attention_heads, encoder_seq_length, encoder_key_length],
) )
@is_flaky() @unittest.skip(
reason="In training model, the first speech encoder layer is sometimes skipped. Training is not supported yet, so the test is ignored."
)
def test_retain_grad_hidden_states_attentions(self): def test_retain_grad_hidden_states_attentions(self):
super().test_retain_grad_hidden_states_attentions() pass
@require_torch @require_torch
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment