Wrong format?

#1
by AlexPoto - opened

LM-Studio did not recognise this model as VL, unlike previous models.

Any update on this?

FYI, both llama-llava-cli and llama-cli make it produce gibberish:

llama-cli -c 0  --temp 0.2 -m Qwen2.5-VL-7B-Instruct-Q4_K_M.gguf  -p "Provide a full description." 
...
<|im_start|>system
You are a helpful assistant<|im_end|>
<|im_start|>user
Hello<|im_end|>
<|im_start|>assistant
Hi there<|im_end|>
<|im_start|>user
How are you?<|im_end|>
<|im_start|>assistant


system_info: n_threads = 6 (n_threads_batch = 6) / 12 | CPU : SSE3 = 1 | SSSE3 = 1 | AVX = 1 | AVX2 = 1 | F16C = 1 | FMA = 1 | LLAMAFILE = 1 | OPENMP = 1 | AARCH64_REPACK = 1 | 

main: interactive mode on.
sampler seed: 1971700995
sampler params: 
    repeat_last_n = 64, repeat_penalty = 1,000, frequency_penalty = 0,000, presence_penalty = 0,000
    dry_multiplier = 0,000, dry_base = 1,750, dry_allowed_length = 2, dry_penalty_last_n = 128000
    top_k = 40, top_p = 0,950, min_p = 0,050, xtc_probability = 0,000, xtc_threshold = 0,100, typical_p = 1,000, temp = 0,200
    mirostat = 0, mirostat_lr = 0,100, mirostat_ent = 5,000
sampler chain: logits -> logit-bias -> penalties -> dry -> top-k -> typical -> top-p -> min-p -> xtc -> temp-ext -> dist 
generate: n_ctx = 128000, n_batch = 2048, n_predict = -1, n_keep = 0

== Running in interactive mode. ==
 - Press Ctrl+C to interject at any time.
 - Press Return to return control to the AI.
 - To return control without starting a new line, end your input with '/'.
 - If you want to submit another line, end your input with '\'.

system
Provide a full description.

> Describe yourself and your goals
 1141:14:44:14:114 *1:41:44 *1:4 A4 A4:1:4 A. *4: * A: * A4:4 * * A: * * * * * A:4: * * * * * * * * * * A:4 * A * * 1: * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * A * * * * * A * * * * A * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * A * * * * * * * * * * * A * * * * * * A * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * A * * * * * * * * * * * * * * * * * * A * * * * * * * * * * * * * * * * * * * * * * * * * * * * *^C^C^C^C^C^C

Ref, hand-compiled:

llama-cli --version
version: 4591 (7919256c)
built with Ubuntu clang version 18.1.8 

works as a regular language model but no vision support(LM Studio)

Do we need a mmproj file to enable the vision? Was this split out of the model when converted to gguf?

For me both my bot this quant generate gibberish.
I assume the vision adapter extractor (..._surgery...py) is not implemented (or not correctly) in llama.cpp (not even tried as the language part not working - at least for me -)
so the vision adapters not extracted.

Sign up or log in to comment