1010from ...utils import check_embeddings_close , check_transformers_version
1111
1212
13- @pytest .fixture (autouse = True )
14- def v1 (run_with_both_engines ):
15- # Simple autouse wrapper to run both engines for each test
16- # This can be promoted up to conftest.py to run for every
17- # test in a package
18- pass
19-
20-
2113@pytest .mark .parametrize (
2214 "model" ,
2315 [
@@ -32,21 +24,15 @@ def v1(run_with_both_engines):
3224 "intfloat/e5-mistral-7b-instruct" ,
3325 # CPU v1 doesn't support sliding window
3426 marks = [pytest .mark .core_model ]),
35- # the qwen models interfere with each other (see PR
36- # https://github.com/vllm-project/vllm/pull/18720).
37- # To avoid this problem, for now we skip v0 since it will be
38- # deprecated anyway.
3927 pytest .param ("ssmits/Qwen2-7B-Instruct-embed-base" ,
40- marks = [pytest .mark .skip_v0 , pytest . mark . cpu_model ]),
28+ marks = [pytest .mark .cpu_model ]),
4129 # [Encoder-only]
4230 pytest .param ("BAAI/bge-base-en-v1.5" , marks = [pytest .mark .core_model ]),
4331 pytest .param ("sentence-transformers/all-MiniLM-L12-v2" ),
4432 pytest .param ("intfloat/multilingual-e5-small" ),
45- pytest .param ("Alibaba-NLP/gte-Qwen2-1.5B-instruct" ,
46- marks = [pytest .mark .skip_v1 ]),
33+ pytest .param ("Alibaba-NLP/gte-Qwen2-1.5B-instruct" ),
4734 # [Cross-Encoder]
48- pytest .param ("sentence-transformers/stsb-roberta-base-v2" ,
49- marks = [pytest .mark .skip_v1 ]),
35+ pytest .param ("sentence-transformers/stsb-roberta-base-v2" ),
5036 ],
5137)
5238def test_models (
0 commit comments