holunda-io / bpm-ai-connectors-camunda-8

🤖 Boost automation in your Camunda processes using task-specific AI connectors
Apache License 2.0
32 stars 2 forks source link

Cache predictions #154

Closed Benjoyo closed 5 months ago

Benjoyo commented 5 months ago

To improve latency and costs, LLM completions, OCR results and other predictions should be cached (disk backed). Taking into account all kinds of parameters that should invalidate (different model or changed prompt) or disable (e.g. temperature > 0) a cache.