mirror of
https://github.com/huggingface/transformers.git
synced 2025-10-20 09:03:53 +08:00
@ -242,10 +242,6 @@ class MPNetModelTest(ModelTesterMixin, PipelineTesterMixin, unittest.TestCase):
|
||||
config_and_inputs = self.model_tester.prepare_config_and_inputs()
|
||||
self.model_tester.create_and_check_mpnet_for_question_answering(*config_and_inputs)
|
||||
|
||||
@unittest.skip(reason="TFMPNet adds poolers to all models, unlike the PT model class.")
|
||||
def test_tf_from_pt_safetensors(self):
|
||||
return
|
||||
|
||||
|
||||
@require_torch
|
||||
class MPNetModelIntegrationTest(unittest.TestCase):
|
||||
|
@ -600,10 +600,6 @@ class Speech2TextModelTest(ModelTesterMixin, GenerationTesterMixin, PipelineTest
|
||||
def test_generate_without_input_ids(self):
|
||||
pass
|
||||
|
||||
@unittest.skip(reason="Test failing, @RocketNight is looking into it")
|
||||
def test_tf_from_pt_safetensors(self):
|
||||
pass
|
||||
|
||||
|
||||
@require_torch
|
||||
@require_torchaudio
|
||||
|
@ -520,10 +520,6 @@ class TapasModelTest(ModelTesterMixin, PipelineTesterMixin, unittest.TestCase):
|
||||
config_and_inputs = self.model_tester.prepare_config_and_inputs()
|
||||
self.model_tester.create_and_check_for_sequence_classification(*config_and_inputs)
|
||||
|
||||
@unittest.skip(reason="tfp is not defined even if installed. FIXME @Arthur in a followup PR!")
|
||||
def test_tf_from_pt_safetensors(self):
|
||||
pass
|
||||
|
||||
|
||||
def prepare_tapas_single_inputs_for_inference():
|
||||
# Here we prepare a single table-question pair to test TAPAS inference on:
|
||||
|
@ -1344,14 +1344,6 @@ class ModelTesterMixin:
|
||||
[self.model_tester.num_attention_heads, encoder_seq_length, encoder_key_length],
|
||||
)
|
||||
|
||||
# This is copied from `torch/testing/_internal/jit_utils.py::clear_class_registry`
|
||||
def clear_torch_jit_class_registry(self):
|
||||
torch._C._jit_clear_class_registry()
|
||||
torch.jit._recursive.concrete_type_store = torch.jit._recursive.ConcreteTypeStore()
|
||||
# torch 1.8 has no `_clear_class_state` in `torch.jit._state`
|
||||
if hasattr(torch.jit._state, "_clear_class_state"):
|
||||
torch.jit._state._clear_class_state()
|
||||
|
||||
def test_hidden_states_output(self):
|
||||
def check_hidden_states_output(inputs_dict, config, model_class):
|
||||
model = model_class(copy.deepcopy(config))
|
||||
|
Reference in New Issue
Block a user