simonw / llm-replicate

LLM plugin for models hosted on Replicate
Apache License 2.0
58 stars 6 forks source link

Allow replicate model to be used in non-chat mode #16

Open cmungall opened 1 year ago

cmungall commented 1 year ago

llm -m replicate-a16z-infra-llama13b-v2-chat "ten names for a pet pelican"

generates a reasonable response, but if I hop over to my replicate dashboard I see that the actual prompt that was issued is:

User: ten names for a pet pelican
Assistant:

I believe this happens when chat mode is set.

I am finding when I play with llama2 via the replicate dashboard, this form of prompt doesn't work well for my use case (json extraction using examples derived via RAG) - especially for 13b. 70b is a bit better, but the trailing Assistant: still seems to confound it.

I am not sure if this is an issue with this plugin, or with how it's instantiated in the host llm package, but either way I can't figure out how to pass my prompt directly