-
Notifications
You must be signed in to change notification settings - Fork 11.9k
LoRA Model Support and Full Fine-tune options #4645
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
@the-crypt-keeper would you know either answer here? - you seem knowledgeable on the llama.cpp examples. Thanks |
Regarding full fine-tuning, on closer reading, I see that this example allows for full fine-tuning (incl. training from scratch): train-text-from-scratch |
@RonanKMcGovern you found it, |
I tried fine-tuning TinyLlama locally for chat and couldn't even load it - probably was a bad model to pick: #4703 . Easy to do on transformers but would be nice to be able to do it on mac. |
This issue is stale because it has been open for 30 days with no activity. |
This issue was closed because it has been inactive for 14 days since being marked as stale. |
Full fine-tuning with Llama.cpp
Are there any efforts to allow for full-fine tuning with Llama cpp (not LoRA)?
If not, is this because there isn't a back-propagation capability? It would be great to at least be able to train the norm and embed modules so that model context can be extended. This also improves performance when chat fine-tuning models.
LoRA fine-tuning
I notice in the examples that only Llama models are supported, but I see in the issues that there are other models being fine-tuned successfully. Is there a list of what works and what doesn't? Happy to make a PR on the example to update that.
The text was updated successfully, but these errors were encountered: