Skip to content

Conversation

mudler
Copy link
Owner

@mudler mudler commented Mar 13, 2025

Description

Until a better solution is found upstream, be conservative and default to GPU.

ggml-org/llama.cpp#12322
ggml-org/llama.cpp#12322 (comment)

This PR slows things on GPU (clip on gpu makes it to run much faster!), however it breaks non-GPU usage.

To enable clip GPU usage specifically, set gpu in the model options:

options:
- gpu

Notes for Reviewers

Signed commits

  • Yes, I signed my commits.

Until a better solution is found upstream, be conservative and default
to GPU.

ggml-org/llama.cpp#12322
ggml-org/llama.cpp#12322 (comment)

Signed-off-by: Ettore Di Giacinto <[email protected]>
@mudler mudler added the bug Something isn't working label Mar 13, 2025
Copy link

netlify bot commented Mar 13, 2025

Deploy Preview for localai ready!

Name Link
🔨 Latest commit 2c5a308
🔍 Latest deploy log https://app.netlify.com/sites/localai/deploys/67d2ca2206b33b0008f6aaab
😎 Deploy Preview https://deploy-preview-5010--localai.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify site configuration.

@mudler mudler changed the title fix(clip): do not imply GPUs by default fix(clip): do not imply GPU offload by default Mar 13, 2025
@mudler mudler merged commit 423514a into master Mar 13, 2025
25 checks passed
@mudler mudler deleted the fix/clip branch March 13, 2025 14:14
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant