dumping tensors with llama-eval-callback #14567
-
|
I am trying to dump the tensors of a model, its a vlm and I can get the text model tensors but there doesn't seem to be a --mmproj argument and was wondering if this exists elsewhere. |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
|
As far as I am aware, If you're working with a VLM (e.g. Qwen-VL, BLIP-style models), the
To access
Currently, there is no CLI argument equivalent to Would love to work on a feature regarding this, let me know if any help is required. |
Beta Was this translation helpful? Give feedback.
As far as I am aware,
--mmprojdoes not currently exist as a command-line argument inllama-eval-callbackor elsewhere inllama.cpp. The model loader inllama.cppis primarily designed around text-only models like LLaMA, and while there is emerging support for vision-language models (VLMs), the integration of vision-specific components likemm_projisn't fully exposed through the CLI yet.If you're working with a VLM (e.g. Qwen-VL, BLIP-style models), the
mm_projtensor typically corresponds to a linear projection layer used to align visual embeddings with the language model. These weights may exist in the original model files (e.g., in Hugging Facesafetensors), but during conversion to.…