-
Notifications
You must be signed in to change notification settings - Fork 11.8k
server: Describing pictures with multi models seems to crash the model #13480
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
I have the same issue whit Rolmocr model that is a fine-tuned Qwen2.5 model. Only text prompt is fonctionnning but when I send an image with text there is a crash... ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no |
Just tried at home with build 5359 and this model: |
Works as well here in the office. |
Ok, I don't understand with they are so much memory allocated in the buffer when the model receive the image... I hope that the last build fix this issue. |
I confirm, it's working with the latest build. |
Hi all,
Tried to describe a picture with these two models in separate runs:
https://huggingface.co/bartowski/mistralai_Mistral-Small-3.1-24B-Instruct-2503-GGUF
https://huggingface.co/bartowski/Qwen_Qwen2.5-VL-7B-Instruct-GGUF
The llama.cpp build used was b5351 CPU X64 on Win 11.
No errors where thrown.
Greetings,
Simon
The text was updated successfully, but these errors were encountered: