You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, I'm running some offline inference benchmarks using llama-cpp-python, and the prompt cache that was implemented here (#158) is getting in the way of measuring prompt evaluation time. Is there an option to disable it?
The text was updated successfully, but these errors were encountered:
Hi, I'm running some offline inference benchmarks using llama-cpp-python, and the prompt cache that was implemented here (#158) is getting in the way of measuring prompt evaluation time. Is there an option to disable it?
The text was updated successfully, but these errors were encountered: