Replies: 1 comment 11 replies
-
It's up to the clients to support it. On our side, we can add an option to enable it by default for all requests. Maybe it's not a bad idea since I think it's always better to have it enabled. |
Beta Was this translation helpful? Give feedback.
11 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
How are you guys effectively using the cache_prompt parameter? I am using my local LLM with openweb ui and aider, but AFAIK neither has the ability to set the cache_prompt parameter in their requests. So how are you guys effectively enabling the cache with so many clients not supporting this parameter?
Is there some other way to enable caching? The prompt processing would be a lot faster if earlier parts of our conversation could be cached.
Beta Was this translation helpful? Give feedback.
All reactions