-
Notifications
You must be signed in to change notification settings - Fork 11.6k
ggml_new_tensor_impl: not enough space in the context's memory pool #29
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Tried out #31 - it, uh, got farther: |
ok I did an upsie in that PR, initializing it that way apparently didn't zero'ed out the rest of the fields. I updated the branch, please test it again now! |
It started to expand the prompt, but with seemingly garbage data: |
Should be good on latest master - reopen if issue persists. |
Hey i was trying to run this on a RHEL 8 server with 32 cpu cores. and i am getting the same error. On my second query. I am using GPT4All-J v1.3-groovy.
|
Hi @ggerganov @gjmulder I would appreciate some direction for this pls. |
Getting the same issue on Apple M1 Pro with 16GB RAM when trying the example from: Using a relatively large PDF with ~200 pages Stack trace: gpt_tokenize: unknown token '?' |
Same issue when running on Win11 with 64GB RAM (25 GB utilized): ggml_new_tensor_impl: not enough space in the scratch memory pool (needed 450887680, available 446693376) |
Oh hey, exact same error:
|
Same issue here, tried a combination of settings but just keep getting the memory error even though both RAM and GPU RAM are less than 50% utilization. I had to follow the guide here to build llama-cpp with GPU support as it wasn't working previously, but even before that it was giving the same error (side note GPU support natively does work in oobabooga windows!?): Anyone have any ideas? HW: Using embedded DuckDB with persistence: data will be stored in: db What would you like to know about the policies? test ggml_new_object: not enough space in the context's memory pool (needed 10882896, available 10650320) |
Same here... any solutions already??? |
Solved this by going back to llama-cpp-python version 0.1.74 |
well this has nothing to do with python |
Same here... any solutions already??? |
@dereklll This issue was closed 6 months ago, I'd suggest to create a new one. |
Same issue on a runpod gpu machine, tried 2 different gpu's |
Fixes and Tweaks to Defaults
Heya! Friend showed this to me and I'm trying to get it to work myself on Windows 10. I've applied the changes as seen in #22 to get it to build (more specifically, I pulled in the new commits from etra0's fork, but the actual executable fails to run - printing this before segfaulting:
I'm trying to use 7B on an i9-13900K (and I have about 30 gigs of memory free right now), and I've verified my hashes with a friend. Any ideas? Thanks!
The text was updated successfully, but these errors were encountered: