GGUF conversion fails

#3
by Lokator - opened

Hello!
I'm trying to use VILA / NVILA / LongVILA models in Ollama + OpenWebUI.

My problem is that when trying to use e.g ggml-org/gguf-my-repo or other GGUF converters, they always fail;
the error traces back to Llama.cpp backend and it says:

"Error converting to fp16: INFO:hf-to-gguf:Loading model: NVILA-15B
ERROR:hf-to-gguf:Model LlavaLlamaModel is not supported"

Llama.cpp doesn't seem to like 'LlavaLlamaModel' as a base-model.

(basically llama_cpp/convert_hf_to_gguf.py throws a NotImplementedError)

Can you offer any help with this?
I'd really appreciate any tips, or, best case scenario, .gguf files offered under your projects.

Sign up or log in to comment