From 13b2a8cf156011a25073f0c1203196f39bb90645 Mon Sep 17 00:00:00 2001 From: Isotr0py <2037008807@qq.com> Date: Tue, 12 Aug 2025 00:45:48 +0800 Subject: [PATCH] skip mllama test Signed-off-by: Isotr0py <2037008807@qq.com> --- .../models/multimodal/generation/test_mllama.py | 17 +++++++++++++++++ 1 file changed, 17 insertions(+) diff --git a/tests/models/multimodal/generation/test_mllama.py b/tests/models/multimodal/generation/test_mllama.py index 2bb01e494d43..b413c4d6b366 100644 --- a/tests/models/multimodal/generation/test_mllama.py +++ b/tests/models/multimodal/generation/test_mllama.py @@ -6,6 +6,7 @@ import pytest import torch from transformers import AutoConfig, AutoModelForImageTextToText, AutoTokenizer +from transformers import __version__ as TRANSFORMERS_VERSION from vllm import LLM, SamplingParams from vllm.attention.backends.flash_attn import FlashAttentionMetadata @@ -285,6 +286,10 @@ def clear_cache(): @pytest.mark.parametrize("max_tokens", [128]) @pytest.mark.parametrize("num_logprobs", [5]) @pytest.mark.parametrize("attn_backend", LIST_ENC_DEC_SUPPORTED_BACKENDS) +@pytest.mark.skipif( + TRANSFORMERS_VERSION == "4.55.0", + reason="Transformers v4.55.0 has a regression issue on mllama, " + "see: https://github.com/huggingface/transformers/pull/40083") def test_models_single_leading_image(hf_runner, vllm_runner, image_assets, model, sizes, dtype, max_tokens, num_logprobs, @@ -313,6 +318,10 @@ def test_models_single_leading_image(hf_runner, vllm_runner, image_assets, @pytest.mark.parametrize("max_tokens", [128]) @pytest.mark.parametrize("num_logprobs", [5]) @pytest.mark.parametrize("attn_backend", LIST_ENC_DEC_SUPPORTED_BACKENDS) +@pytest.mark.skipif( + TRANSFORMERS_VERSION == "4.55.0", + reason="Transformers v4.55.0 has a regression issue on mllama, " + "see: https://github.com/huggingface/transformers/pull/40083") def test_models_multi_leading_images(hf_runner, vllm_runner, image_assets, model, dtype, max_tokens, num_logprobs, attn_backend: _Backend) -> None: @@ -362,6 +371,10 @@ def test_models_multi_leading_images(hf_runner, vllm_runner, image_assets, @pytest.mark.parametrize("max_tokens", [128]) @pytest.mark.parametrize("num_logprobs", [5]) @pytest.mark.parametrize("attn_backend", LIST_ENC_DEC_SUPPORTED_BACKENDS) +@pytest.mark.skipif( + TRANSFORMERS_VERSION == "4.55.0", + reason="Transformers v4.55.0 has a regression issue on mllama, " + "see: https://github.com/huggingface/transformers/pull/40083") def test_models_interleaved_images(hf_runner, vllm_runner, image_assets, model, dtype, max_tokens, num_logprobs, attn_backend: _Backend) -> None: @@ -402,6 +415,10 @@ def test_models_interleaved_images(hf_runner, vllm_runner, image_assets, model, @pytest.mark.parametrize("dtype", ["bfloat16"]) @pytest.mark.parametrize("max_tokens", [64]) @pytest.mark.parametrize("num_logprobs", [5]) +@pytest.mark.skipif( + TRANSFORMERS_VERSION == "4.55.0", + reason="Transformers v4.55.0 has a regression issue on mllama, " + "see: https://github.com/huggingface/transformers/pull/40083") def test_models_distributed( hf_runner, vllm_runner,