List of supported models (and sticky)? #14601
Replies: 3 comments
-
type |
Beta Was this translation helpful? Give feedback.
-
Not quite. Some seem to be, some not. Not format wise (gguf is there), but function wise - tokenization doesn't work. Qwen3 Embedding and Qwen3 Reranking models are not supported yet. Also the GGUFs provided by Qwen are missing important metadata and will not work until fixed and supported. |
Beta Was this translation helpful? Give feedback.
-
https://github.com/ggml-org/llama.cpp/blob/master/src/llama-arch.h will show you supported architectures, mtmd is being worked towards so vlms wont be immeditaly obvious, finding embeddings ( or tensor names ) you can use the llama-eval-callback script to dump the tensors and to generate gguf from hf you would use the convert_hf_to_gguf to generate .gguf files from their hf impl |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Where can I find list of supported model?
Embedding: ?
Rerank: ?
Generate: ?
Thx
Beta Was this translation helpful? Give feedback.
All reactions