Skip to content

Segfaults now with latest llama.cpp commits #727

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
Ph0rk0z opened this issue Sep 17, 2023 · 4 comments
Closed

Segfaults now with latest llama.cpp commits #727

Ph0rk0z opened this issue Sep 17, 2023 · 4 comments

Comments

@Ph0rk0z
Copy link

Ph0rk0z commented Sep 17, 2023

I pulled new commits for vendor/llama.cpp and I get terminate after throwing an instance of std::length error

The model loads fine and everything seems normal, just dies when generating. This is using the built in server.
Same error is in textgen.

@manishshettym
Copy link

manishshettym commented Sep 18, 2023

Can reproduce! Here's a gdb back trace that shows the issue:

# 5 0x00007fffb935a209 in std::__throw_length_error(char const*) ()
at /opt/conda/conda-bld/gcc-compiler_1654084175708/work/gcc/libstdc++-v3/src/c++98/bitmap_allocator.cc:92
No symbol table info available.
...
# 9 0x00007fffb966b446 in llama_tokenize (ctx=0xfa66e0,
text=0x7ffff6ce9c80 "Hello, world!", **text_len=-153556112**, tokens=0xd,
n_max_tokens=0, add_bos=111)
at ....

It is due to a recent update to the tokenize APIs in llama.cpp.

Opened this PR that fixes it: #730 [cc: @abetlen]

@Ph0rk0z
Copy link
Author

Ph0rk0z commented Sep 18, 2023

Makes sense, that's a big component.

@abetlen
Copy link
Owner

abetlen commented Sep 18, 2023

@Ph0rk0z were you using the llama.cpp version pinned in the vendor/llama.cpp directory?

@Ph0rk0z
Copy link
Author

Ph0rk0z commented Sep 18, 2023

nope, I like to do the latest and greatest

@Ph0rk0z Ph0rk0z closed this as completed Sep 19, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants