-
-
Notifications
You must be signed in to change notification settings - Fork 2.7k
Closed
Labels
area/backendsenhancementNew feature or requestNew feature or requestroadmapup for grabsTickets that no-one is currently working onTickets that no-one is currently working on
Description
Is your feature request related to a problem? Please describe.
Hello, I tried ollama
on my macbook and got pretty good performance compared to running LocalAI
with llama-stable
directly(which consumes lots of CPU and not using GPU at all):

While Ollama
will use the GPU and so saves CPU, but unfortunately ollama
did not have OpenAI like API.
Describe the solution you'd like
Add support for ollama.
Describe alternatives you've considered
Had not found one proper.
Additional context
Thanks a lot :D
gobetti and lboklin
Metadata
Metadata
Assignees
Labels
area/backendsenhancementNew feature or requestNew feature or requestroadmapup for grabsTickets that no-one is currently working onTickets that no-one is currently working on