@@ -65,54 +65,54 @@ subprojects:
6565 - " engine-unit-test-PR-test"
6666 - " Genreate-Engine-Report"
6767
68- - id : " Windows Binary Test"
69- paths :
70- - " .github/workflows/windows-test.yml"
71- - " requirements.txt"
72- - " setup.py"
73- - " intel_extension_for_transformers/transformers/runtime/**"
74- - " intel_extension_for_transformers/transformers/llm/operator/**"
75- - " !intel_extension_for_transformers/transformers/runtime/third_party/**"
76- - " !intel_extension_for_transformers/transformers/runtime/docs/**"
77- - " !intel_extension_for_transformers/transformers/runtime/test/**"
78- checks :
79- - " Windows-Binary-Test"
68+ # - id: "Windows Binary Test"
69+ # paths:
70+ # - ".github/workflows/windows-test.yml"
71+ # - "requirements.txt"
72+ # - "setup.py"
73+ # - "intel_extension_for_transformers/transformers/runtime/**"
74+ # - "intel_extension_for_transformers/transformers/llm/operator/**"
75+ # - "!intel_extension_for_transformers/transformers/runtime/third_party/**"
76+ # - "!intel_extension_for_transformers/transformers/runtime/docs/**"
77+ # - "!intel_extension_for_transformers/transformers/runtime/test/**"
78+ # checks:
79+ # - "Windows-Binary-Test"
8080
81- - id : " LLM Model Test workflow"
82- paths :
83- - " .github/workflows/llm-test.yml"
84- - " .github/workflows/script/models/run_llm.sh"
85- - " intel_extension_for_transformers/transformers/runtime/**"
86- - " !intel_extension_for_transformers/transformers/runtime/kernels/**"
87- - " !intel_extension_for_transformers/transformers/runtime/test/**"
88- - " !intel_extension_for_transformers/transformers/runtime/third_party/**"
89- - " !intel_extension_for_transformers/transformers/runtime/docs/**"
90- checks :
91- - " LLM-Workflow (gpt-j-6b, engine, latency, bf16,int8,fp8)"
92- - " Generate-LLM-Report"
81+ # - id: "LLM Model Test workflow"
82+ # paths:
83+ # - ".github/workflows/llm-test.yml"
84+ # - ".github/workflows/script/models/run_llm.sh"
85+ # - "intel_extension_for_transformers/transformers/runtime/**"
86+ # - "!intel_extension_for_transformers/transformers/runtime/kernels/**"
87+ # - "!intel_extension_for_transformers/transformers/runtime/test/**"
88+ # - "!intel_extension_for_transformers/transformers/runtime/third_party/**"
89+ # - "!intel_extension_for_transformers/transformers/runtime/docs/**"
90+ # checks:
91+ # - "LLM-Workflow (gpt-j-6b, engine, latency, bf16,int8,fp8)"
92+ # - "Generate-LLM-Report"
9393
94- - id : " Chat Bot Test workflow"
95- paths :
96- - " .github/workflows/chatbot-test.yml"
97- - " .github/workflows/chatbot-inference-llama-2-7b-chat-hf.yml"
98- - " .github/workflows/chatbot-inference-mpt-7b-chat.yml"
99- - " .github/workflows/chatbot-finetune-mpt-7b-chat.yml"
100- - " .github/workflows/chatbot-inference-llama-2-7b-chat-hf-hpu.yml"
101- - " .github/workflows/chatbot-inference-mpt-7b-chat-hpu.yml"
102- - " .github/workflows/chatbot-finetune-mpt-7b-chat-hpu.yml"
103- - " .github/workflows/script/chatbot/**"
104- - " .github/workflows/sample_data/**"
105- - " intel_extension_for_transformers/neural_chat/**"
106- - " intel_extension_for_transformers/transformers/llm/finetuning/**"
107- - " intel_extension_for_transformers/transformers/llm/quantization/**"
108- - " intel_extension_for_transformers/transformers/**"
109- - " workflows/chatbot/inference/**"
110- - " workflows/chatbot/fine_tuning/**"
111- - " !intel_extension_for_transformers/neural_chat/docs/**"
112- - " !intel_extension_for_transformers/neural_chat/tests/ci/**"
113- - " !intel_extension_for_transformers/neural_chat/examples/**"
114- - " !intel_extension_for_transformers/neural_chat/assets/**"
115- - " !intel_extension_for_transformers/neural_chat/README.md"
116- checks :
117- - " call-inference-llama-2-7b-chat-hf / inference test"
118- - " call-inference-mpt-7b-chat / inference test"
94+ # - id: "Chat Bot Test workflow"
95+ # paths:
96+ # - ".github/workflows/chatbot-test.yml"
97+ # - ".github/workflows/chatbot-inference-llama-2-7b-chat-hf.yml"
98+ # - ".github/workflows/chatbot-inference-mpt-7b-chat.yml"
99+ # - ".github/workflows/chatbot-finetune-mpt-7b-chat.yml"
100+ # - ".github/workflows/chatbot-inference-llama-2-7b-chat-hf-hpu.yml"
101+ # - ".github/workflows/chatbot-inference-mpt-7b-chat-hpu.yml"
102+ # - ".github/workflows/chatbot-finetune-mpt-7b-chat-hpu.yml"
103+ # - ".github/workflows/script/chatbot/**"
104+ # - ".github/workflows/sample_data/**"
105+ # - "intel_extension_for_transformers/neural_chat/**"
106+ # - "intel_extension_for_transformers/transformers/llm/finetuning/**"
107+ # - "intel_extension_for_transformers/transformers/llm/quantization/**"
108+ # - "intel_extension_for_transformers/transformers/**"
109+ # - "workflows/chatbot/inference/**"
110+ # - "workflows/chatbot/fine_tuning/**"
111+ # - "!intel_extension_for_transformers/neural_chat/docs/**"
112+ # - "!intel_extension_for_transformers/neural_chat/tests/ci/**"
113+ # - "!intel_extension_for_transformers/neural_chat/examples/**"
114+ # - "!intel_extension_for_transformers/neural_chat/assets/**"
115+ # - "!intel_extension_for_transformers/neural_chat/README.md"
116+ # checks:
117+ # - "call-inference-llama-2-7b-chat-hf / inference test"
118+ # - "call-inference-mpt-7b-chat / inference test"
0 commit comments