Closed hamzaziizzz closed 2 months ago
are you able to integrate it?
Hey @hamzaziizzz and @SanjuktaK,
In the Rasa Pro 3.10 release, we've improved the configuration process for self-hosted LLMs. To use a self-hosted LLM, you need to set provider: self-hosted
and specify the correct URL for your OpenAI-compatible API endpoint using the api_base
key. Here’s an example configuration:
- name: SingleStepLLMCommandGenerator
llm:
provider: self-hosted
model: meta-llama/CodeLlama-7b-Instruct-hf
api_base: "https://my-endpoint/v1"
Make sure you're using Rasa Pro version 3.10 or higher. You can find more details in our documentation: Rasa Pro - LLM Configuration - Self Hosted Model Server.
As for the Llama model itself, I recently tested one in the GGUF format, and it worked seamlessly with CALM. Here’s the configuration I used:
- name: SingleStepLLMCommandGenerator
llm:
provider: "self-hosted"
model: "ggml-org/Meta-Llama-3.1-8B-Instruct-Q4_0-GGUF"
api_base: "http://localhost:8080/v1"
request_timeout: 30
temperature: 0.0
top_p: 0.0
I hope this helps.
I am geeting the following for Llama-LLM
My
config.yml
is as follows