You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
As far as I can tell the ggcc files are not compatible with llama.cpp and are instead used by a fork?
It looks like the mainline llama.cpp does not yet support Falcon. Here are some relevant upsteam issue to keep an eye on, once these are completed we should be able to add Falcon support:
GGCC is for https://github.com/cmp-nct/ggllm.cpp
ggllm is outdated, it is quite a bit higher featured than llama.cpp for Falcon, also faster in generation and likely uses less memory for the same job. But it's much slower in batched processing (prompt ingestion).
https://huggingface.co/TheBloke/falcon-7b-instruct-GGML/tree/main
I encounter this error when I try to load the 5bit model:
The text was updated successfully, but these errors were encountered: