Skip to content

Thread Safety in llama.cpp #596

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
martindevans opened this issue Mar 12, 2024 · 4 comments
Open

Thread Safety in llama.cpp #596

martindevans opened this issue Mar 12, 2024 · 4 comments
Labels
do not close Protect this issue from auto closing Upstream Tracking an issue in llama.cpp

Comments

@martindevans
Copy link
Member

Tracking issue for thread safety in llama.cpp. The global inference lock can be removed once this is resolved.

ggml-org/llama.cpp#3960

@martindevans martindevans added the Upstream Tracking an issue in llama.cpp label Mar 12, 2024
@zsogitbe
Copy link
Contributor

llama.cpp : add pipeline parallelism support #6017. Good news: seems high priority and will probably be ready soon. If this and the CUDA memory release bug correction is ready please add a quick intermediate release integration to LLamaSharp. This is important.

ggml-org/llama.cpp#6017

@gospelask
Copy link

Great! Great! Great!

Copy link

github-actions bot commented May 6, 2025

This issue has been automatically marked as stale due to inactivity. If no further activity occurs, it will be closed in 7 days.

@github-actions github-actions bot added the stale Stale issue will be autoclosed soon label May 6, 2025
@martindevans martindevans added do not close Protect this issue from auto closing and removed stale Stale issue will be autoclosed soon labels May 6, 2025
@martindevans
Copy link
Member Author

Now that a lot of time has passed this should be reviewed again. It would be nice to remove the GlobalInferenceLock that LLamaSharp currently has (see here)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
do not close Protect this issue from auto closing Upstream Tracking an issue in llama.cpp
Projects
None yet
Development

No branches or pull requests

3 participants