zilliztech / GPTCache

Semantic cache for LLMs. Fully integrated with LangChain and llama_index.
https://gptcache.readthedocs.io
MIT License
6.89k stars 480 forks source link

[Feature]: GPTCache implementation for GPT Assistants #573

Open yashaswi-nayak opened 7 months ago

yashaswi-nayak commented 7 months ago

Is your feature request related to a problem? Please describe.

We are creating multiple GPT Assistants to support our customer service operations. Currently, there is only support for caching results via ChatCompletion. We would like to see a GPT Assistant implementation of GPTCache.

GPT Assistants will mainly answer multiple queries that are very common like

Over the course of time, we build a decent cache storage of all answers for user queries. These answers can be pulled from the cache directly instead of requesting Open AI every time. This would save us a lot of tokens and latency time.

Describe the solution you'd like.

I would like to see a solution where we have

Describe an alternate solution.

No response

Anything else? (Additional Context)

No response