@@ -69,11 +69,13 @@ Run local LLMs on iGPU, APU and CPU (AMD , Intel, and Qualcomm (Coming Soon)). E
6969 - ** CPU:** ` $env:ELLM_TARGET_DEVICE='cpu'; pip install -e .[cpu] `
7070 - ** CUDA:** ` $env:ELLM_TARGET_DEVICE='cuda'; pip install -e .[cuda] `
7171 - ** IPEX:** ` $env:ELLM_TARGET_DEVICE='ipex'; python setup.py develop `
72+ - ** OpenVINO:** ` $env:ELLM_TARGET_DEVICE='openvino'; pip install -e .[openvino] `
7273 - ** With Web UI** :
7374 - ** DirectML:** ` $env:ELLM_TARGET_DEVICE='directml'; pip install -e .[directml,webui] `
7475 - ** CPU:** ` $env:ELLM_TARGET_DEVICE='cpu'; pip install -e .[cpu,webui] `
7576 - ** CUDA:** ` $env:ELLM_TARGET_DEVICE='cuda'; pip install -e .[cuda,webui] `
7677 - ** IPEX:** ` $env:ELLM_TARGET_DEVICE='ipex'; python setup.py develop; pip install -r requirements-webui.txt `
78+ - ** OpenVINO:** ` $env:ELLM_TARGET_DEVICE='openvino'; pip install -e .[openvino,webui] `
7779
7880- ** Linux**
7981
@@ -88,11 +90,13 @@ Run local LLMs on iGPU, APU and CPU (AMD , Intel, and Qualcomm (Coming Soon)). E
8890 - ** CPU:** ` ELLM_TARGET_DEVICE='cpu' pip install -e .[cpu] `
8991 - ** CUDA:** ` ELLM_TARGET_DEVICE='cuda' pip install -e .[cuda] `
9092 - ** IPEX:** ` ELLM_TARGET_DEVICE='ipex' python setup.py develop `
93+ - ** OpenVINO:** ` ELLM_TARGET_DEVICE='openvino' pip install -e .[openvino] `
9194 - ** With Web UI** :
9295 - ** DirectML:** ` ELLM_TARGET_DEVICE='directml' pip install -e .[directml,webui] `
9396 - ** CPU:** ` ELLM_TARGET_DEVICE='cpu' pip install -e .[cpu,webui] `
9497 - ** CUDA:** ` ELLM_TARGET_DEVICE='cuda' pip install -e .[cuda,webui] `
95- - ** IPEX:** ` $env:ELLM_TARGET_DEVICE='ipex'; python setup.py develop; pip install -r requirements-webui.txt `
98+ - ** IPEX:** ` ELLM_TARGET_DEVICE='ipex' python setup.py develop; pip install -r requirements-webui.txt `
99+ - ** OpenVINO:** ` ELLM_TARGET_DEVICE='openvino' pip install -e .[openvino,webui] `
96100
97101### Launch OpenAI API Compatible Server
98102
@@ -131,12 +135,29 @@ It is an interface that allows you to download and deploy OpenAI API compatible
131135
132136## Compile OpenAI-API Compatible Server into Windows Executable
133137
138+ **NOTE:** OpenVINO packaging currently uses `torch==2.4.0`. It will not be able to run due to missing dependencies which is `libomp`. Make sure to install `libomp` and add the `libomp-xxxxxxx.dll` to `C:\\Windows\\System32`.
139+
1341401. Install `embeddedllm`.
1351412. Install PyInstaller: `pip install pyinstaller==6.9.0`.
1361423. Compile Windows Executable: `pyinstaller .\ellm_api_server.spec`.
1371434. You can find the executable in the `dist\ellm_api_server`.
1381445. Use it like `ellm_server`. `.\ellm_api_server.exe --model_path <path/to/model/weight>`.
139145
146+ _Powershell/Terminal Usage_:
147+
148+ ```powershell
149+ ellm_server --model_path <path/to/model/weight>
150+
151+ # DirectML
152+ ellm_server --model_path 'EmbeddedLLM_Phi-3-mini-4k-instruct-062024-onnx\onnx\directml\Phi-3-mini-4k-instruct-062024-int4' --port 5555
153+
154+ # IPEX-LLM
155+ ellm_server --model_path '.\meta-llama_Meta-Llama-3.1-8B-Instruct\' --backend 'ipex' --device 'xpu' --port 5555 --served_model_name 'meta-llama_Meta/Llama-3.1-8B-Instruct'
156+
157+ # OpenVINO
158+ ellm_server --model_path '.\meta-llama_Meta-Llama-3.1-8B-Instruct\' --backend 'openvino' --device 'gpu' --port 5555 --served_model_name 'meta-llama_Meta/Llama-3.1-8B-Instruct'
159+ ```
160+
140161## Prebuilt OpenAI API Compatible Windows Executable (Alpha)
141162
142163You can find the prebuilt OpenAI API Compatible Windows Executable in the Release page.
@@ -151,6 +172,9 @@ _Powershell/Terminal Usage (Use it like `ellm_server`)_:
151172
152173# IPEX-LLM
153174.\ellm_api_server.exe --model_path '.\meta-llama_Meta-Llama-3.1-8B-Instruct\' --backend 'ipex' --device 'xpu' --port 5555 --served_model_name 'meta-llama_Meta/Llama-3.1-8B-Instruct'
175+
176+ # OpenVINO
177+ .\ellm_api_server.exe --model_path '.\meta-llama_Meta-Llama-3.1-8B-Instruct\' --backend 'openvino' --device 'gpu' --port 5555 --served_model_name 'meta-llama_Meta/Llama-3.1-8B-Instruct'
154178```
155179
156180## Acknowledgements
0 commit comments