How to load the model?

#1
by Blithe8002 - opened

I want to use the 4bit quantized model to run an OCR related task on an image. Can you please tell me how to do that? I tried llama.cpp but I was not able to make it work.

csabakecskemeti pinned discussion
DevQuasar org
β€’
edited Sep 17

Thanks for raising the issue
The vision adapter has missed the 'mmproj' prefix so that could cause issue, that like the llama.cpp server were not automatically recognizing the vision adapter as they rely on filename patterns.
Filname fixed
Easy way to use is with LMStudio, but it should work similarly with llama.cpp

Sign up or log in to comment