Closed austintlee closed 1 year ago
All of the parameters mentioned above are optional.
cc @ylwu-amzn
We would like to show a disclaimer that this has only been tested with OpenAI and Bedrock Anthropic Claude models. Can we still add this for the 2.11 release?
What do you want to do?
Tell us about your request. Provide a summary of the request and all versions that are affected.
Change 1: Setting up the pipeline
to
New parameters
system_prompt: this is a message sent to LLMs (e.g. OpenAI) as a "system" role.
user_instructions: this is an additional message sent to LLMs as a "user" role. It is not uncommon for user instructions to be sent as a system prompt. This customization allows for experimentation to play with prompts.
Change 2: Using the pipeline
to
New parameters
context_size: this is the number of search results sent to LLMs. This is typically needed to meet the token size limit which can vary model to model, e.g. 4k characters. Alternatively, you can use the
size
search request parameter to control the amount of information sent to LLMs.interaction_size: this is the number of interactions (questions and answers) sent to LLMs. As with the number of search results, this can affect the total number of tokens seen by the LLM. If this is not set, the default interaction size of 10 is used.
timeout: this is the number of seconds the RAG pipeline waits for the remote model (via the connector) to respond. The default timeout is currently 30 seconds.
What other resources are available? Provide links to related issues, POCs, steps for testing, etc.