Skip to content

Commit 6f40350

Browse files
rasmithRandall Smith
andauthored
[CI/Build][AMD] Enable Entrypoints Integration Test (Pooling) to run without error on ROCm (#29212)
Signed-off-by: Randall Smith <ransmith@amd.com> Co-authored-by: Randall Smith <ransmith@amd.com>
1 parent 052950e commit 6f40350

File tree

11 files changed

+66
-0
lines changed

11 files changed

+66
-0
lines changed

tests/entrypoints/pooling/correctness/test_mteb_embed.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,12 @@
1111
run_mteb_embed_task,
1212
)
1313
from tests.utils import RemoteOpenAIServer
14+
from vllm.platforms import current_platform
15+
16+
if current_platform.is_rocm():
17+
pytest.skip(
18+
"Encoder self-attention is not implemented on ROCm.", allow_module_level=True
19+
)
1420

1521
os.environ["VLLM_LOGGING_LEVEL"] = "WARNING"
1622

tests/entrypoints/pooling/correctness/test_mteb_score.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -13,6 +13,12 @@
1313
run_mteb_rerank,
1414
)
1515
from tests.utils import RemoteOpenAIServer
16+
from vllm.platforms import current_platform
17+
18+
if current_platform.is_rocm():
19+
pytest.skip(
20+
"Encoder self-attention is not implemented on ROCm.", allow_module_level=True
21+
)
1622

1723
os.environ["VLLM_LOGGING_LEVEL"] = "WARNING"
1824

tests/entrypoints/pooling/llm/test_embedding.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,12 @@
99

1010
from vllm import LLM, PoolingParams
1111
from vllm.distributed import cleanup_dist_env_and_memory
12+
from vllm.platforms import current_platform
13+
14+
if current_platform.is_rocm():
15+
pytest.skip(
16+
"Encoder self-attention is not implemented on ROCm.", allow_module_level=True
17+
)
1218

1319
MODEL_NAME = "intfloat/multilingual-e5-small"
1420

tests/entrypoints/pooling/llm/test_encode.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,12 @@
77

88
from vllm import LLM, PoolingParams
99
from vllm.distributed import cleanup_dist_env_and_memory
10+
from vllm.platforms import current_platform
11+
12+
if current_platform.is_rocm():
13+
pytest.skip(
14+
"Encoder self-attention is not implemented on ROCm.", allow_module_level=True
15+
)
1016

1117
MODEL_NAME = "intfloat/multilingual-e5-small"
1218

tests/entrypoints/pooling/llm/test_score.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,12 @@
99
from tests.models.utils import softmax
1010
from vllm import LLM, PoolingParams
1111
from vllm.distributed import cleanup_dist_env_and_memory
12+
from vllm.platforms import current_platform
13+
14+
if current_platform.is_rocm():
15+
pytest.skip(
16+
"Encoder self-attention is not implemented on ROCm.", allow_module_level=True
17+
)
1218

1319
MODEL_NAME = "tomaarsen/Qwen3-Reranker-0.6B-seq-cls"
1420

tests/entrypoints/pooling/openai/test_embedding.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -19,6 +19,7 @@
1919
EmbeddingResponse,
2020
PoolingResponse,
2121
)
22+
from vllm.platforms import current_platform
2223
from vllm.transformers_utils.tokenizer import get_tokenizer
2324
from vllm.utils.serial_utils import (
2425
EMBED_DTYPE_TO_TORCH_DTYPE,
@@ -28,6 +29,11 @@
2829
decode_pooling_output,
2930
)
3031

32+
if current_platform.is_rocm():
33+
pytest.skip(
34+
"Encoder self-attention is not implemented on ROCm.", allow_module_level=True
35+
)
36+
3137
MODEL_NAME = "intfloat/multilingual-e5-small"
3238
DUMMY_CHAT_TEMPLATE = """{% for message in messages %}{{message['role'] + ': ' + message['content'] + '\\n'}}{% endfor %}""" # noqa: E501
3339
DTYPE = "bfloat16"

tests/entrypoints/pooling/openai/test_embedding_dimensions.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -12,6 +12,12 @@
1212
from tests.models.utils import EmbedModelInfo
1313
from tests.utils import RemoteOpenAIServer
1414
from vllm.entrypoints.openai.protocol import EmbeddingResponse
15+
from vllm.platforms import current_platform
16+
17+
if current_platform.is_rocm():
18+
pytest.skip(
19+
"Encoder self-attention is not implemented on ROCm.", allow_module_level=True
20+
)
1521

1622
MODELS = [
1723
EmbedModelInfo("intfloat/multilingual-e5-small", is_matryoshka=False),

tests/entrypoints/pooling/openai/test_embedding_long_text.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -16,6 +16,12 @@
1616

1717
from tests.utils import RemoteOpenAIServer
1818
from vllm.entrypoints.openai.protocol import EmbeddingResponse
19+
from vllm.platforms import current_platform
20+
21+
if current_platform.is_rocm():
22+
pytest.skip(
23+
"Encoder self-attention is not implemented on ROCm.", allow_module_level=True
24+
)
1925

2026

2127
def _generate_random_text(word_count: int) -> str:

tests/entrypoints/pooling/openai/test_rerank.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,12 @@
88

99
from tests.utils import RemoteOpenAIServer
1010
from vllm.entrypoints.openai.protocol import PoolingResponse, RerankResponse
11+
from vllm.platforms import current_platform
12+
13+
if current_platform.is_rocm():
14+
pytest.skip(
15+
"Encoder self-attention is not implemented on ROCm.", allow_module_level=True
16+
)
1117

1218
MODEL_NAME = "BAAI/bge-reranker-base"
1319
DTYPE = "bfloat16"

tests/entrypoints/pooling/openai/test_score.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,12 @@
1010

1111
from tests.utils import RemoteOpenAIServer
1212
from vllm.entrypoints.openai.protocol import ScoreResponse
13+
from vllm.platforms import current_platform
14+
15+
if current_platform.is_rocm():
16+
pytest.skip(
17+
"Encoder self-attention is not implemented on ROCm.", allow_module_level=True
18+
)
1319

1420
MODELS = [
1521
{"name": "BAAI/bge-reranker-v2-m3", "is_cross_encoder": True},

0 commit comments

Comments
 (0)