-
Notifications
You must be signed in to change notification settings - Fork 11.9k
converting phi-3-small error. #7922
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
notebook that shows the problem: |
I got past that, convert script works with L3 model but the quantize script stopped working today. Colab. !./llama.cpp/quantize /content/models/"model".gguf "model"-q6_k.gguf q6_k /bin/bash: line 1: ./llama.cpp/quantize: No such file or directory |
that's because now quantize is called llama-quantize. |
How? With the same colab notebook? Did you change something? |
I tried again but I got the same result!
same as before.
|
the same error , Local or https://huggingface.co/spaces/ggml-org/gguf-my-repo |
same problem here, was really hoping this would work with my fine tune as i switch to newer models but no luck |
Phi-3-small is different architecture, see https://huggingface.co/microsoft/Phi-3-small-128k-instruct/blob/main/config.json#L4. Currently, the script supports In other word, it should work with Otherwise, GPT2 series also useful for testing https://huggingface.co/aisuko/gpt2-xl-gguf with 8 CPUs. |
https://huggingface.co/microsoft/Phi-3-small-128k-instruct is still unsupperted. in config.json:
And convert says:
|
@ggerganov please help. |
According to https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard phi-3 is the best for it's size both the small version both the medium version. Too bad is still unsupported. |
This issue was closed because it has been inactive for 14 days since being marked as stale. |
python llama.cpp/convert-hf-to-gguf.py --outtype f16 --outfile /content/Phi-3-small-128k-instruct.f16.gguf /content/Phi-3-small-128k-instruct
The text was updated successfully, but these errors were encountered: