Open edwardsmith999 opened 1 month ago
I'm not familiar with txtai, but I'm pretty sure there would be some more boilerplate required for this to work reliably. Also I'm a little confused by the prompt format. Doesn't seem to be correct for Llama3-instruct?
Thanks @turboderp, txtai
came up as a RAG alternative to llamachain
, discussed in issue #261, but llamachain seems more complex to get working.
Prompt is probably not correct for Llama3, I took the example from the txtai website for a different model and found this form at least returned the expect result. Leaving this here as a code snippet for anyone interested might be best for now.
Using the tutorial here, it seems creating a general class to wrap
ExLlama2
allows it to be used as an LLM for RAG in txtai. I could add this as a file in the example folder (pull request) if useful. Currently, the code below works for me in looking up the most meaningful data item.