If there will be a gguf version?
#1
by
Cran-May
- opened
Now LLaMA.cpp supports Qwen
https://github.com/ggerganov/llama.cpp/pull/4281
Here is a series of qwen models.(The support of 1.8B model is not checked.)
https://huggingface.co/Qwen/Qwen-72B-Chat
https://huggingface.co/Qwen/Qwen-14B-Chat
https://huggingface.co/Qwen/Qwen-7B-Chat
https://huggingface.co/Qwen/Qwen-1_8B-Chat
https://huggingface.co/Qwen/Qwen-Audio-Chat
https://huggingface.co/Qwen/Qwen-VL-Chat
Maybe Internlm? https://github.com/ggerganov/llama.cpp/pull/4283
https://huggingface.co/internlm/internlm-chat-20b
https://huggingface.co/internlm/internlm-chat-7b-v1_1
Qwen-VL Chat seems not to be supported.
This comment has been hidden
This comment has been hidden