Skip to content

Falcon 7b instruct 5bit doesn't work #43

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
ChezzPlaya opened this issue Jul 1, 2023 · 3 comments
Closed

Falcon 7b instruct 5bit doesn't work #43

ChezzPlaya opened this issue Jul 1, 2023 · 3 comments
Labels
question Further information is requested

Comments

@ChezzPlaya
Copy link

https://huggingface.co/TheBloke/falcon-7b-instruct-GGML/tree/main

I encounter this error when I try to load the 5bit model:

error loading model: missing tok_embeddings.weight
@martindevans
Copy link
Member

martindevans commented Aug 6, 2023

As far as I can tell the ggcc files are not compatible with llama.cpp and are instead used by a fork?

It looks like the mainline llama.cpp does not yet support Falcon. Here are some relevant upsteam issue to keep an eye on, once these are completed we should be able to add Falcon support:

@cmp-nct
Copy link

cmp-nct commented Aug 6, 2023

GGCC is for https://github.com/cmp-nct/ggllm.cpp
ggllm is outdated, it is quite a bit higher featured than llama.cpp for Falcon, also faster in generation and likely uses less memory for the same job. But it's much slower in batched processing (prompt ingestion).

@martindevans
Copy link
Member

The tracking issue I mentioned above is now marked as complete. I think the LLamaSharp 0.5.1 release should support Falcon now.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

3 participants