-
Notifications
You must be signed in to change notification settings - Fork 1.1k
Segfaults now with latest llama.cpp commits #727
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Can reproduce! Here's a gdb back trace that shows the issue:
It is due to a recent update to the tokenize APIs in llama.cpp. |
Makes sense, that's a big component. |
@Ph0rk0z were you using the llama.cpp version pinned in the |
nope, I like to do the latest and greatest |
I pulled new commits for vendor/llama.cpp and I get
terminate after throwing an instance of std::length error
The model loads fine and everything seems normal, just dies when generating. This is using the built in server.
Same error is in textgen.
The text was updated successfully, but these errors were encountered: