You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Then you'll need to use a custom chat handler to load the clip model and process the chat messages and images.
512
513
@@ -600,6 +601,100 @@ messages = [
600
601
601
602
</details>
602
603
604
+
<details>
605
+
<summary>Loading a Local Image With Qwen3VL(Thinking/No Thinking)</summary>
606
+
607
+
This script demonstrates how to load a local image, encode it as a base64 Data URI, and pass it to a local Qwen3-VL model (with the 'use_think_prompt' parameter enabled for thinking model, disabled for instruct model) for processing using the llama-cpp-python library.
608
+
609
+
```python
610
+
# Import necessary libraries
611
+
from llama_cpp import Llama
612
+
from llama_cpp.llama_chat_format import Qwen3VLChatHandler
613
+
import base64
614
+
import os
615
+
616
+
# --- Model Configuration ---
617
+
# Define the path to the main model file
618
+
MODEL_PATH=r"./Qwen3-VL-8B-Thinking-F16.gguf"
619
+
# Define the path to the multi-modal projector file
0 commit comments