How to load the model?
#1
pinned
by
Blithe8002
- opened
I want to use the 4bit quantized model to run an OCR related task on an image. Can you please tell me how to do that? I tried llama.cpp but I was not able to make it work.
csabakecskemeti
pinned discussion
Thanks for raising the issue
The vision adapter has missed the 'mmproj' prefix so that could cause issue, that like the llama.cpp server were not automatically recognizing the vision adapter as they rely on filename patterns.
Filname fixed
Easy way to use is with LMStudio, but it should work similarly with llama.cpp