-
**What problem or use case are you trying to solve?**
THis would be ideal to solve any installation issues.
**Describe the UX of the solution you'd like**
Just a working docker-compose.yaml file…
-
This is not a feature/bug, just a question, since there is no discussion board I'll leave it here.
I have multiple hugging face models hosted and running OpenAI compatible endpoints.
After hookin…
-
Is there a possibility that this could use open source LLM?
OpenAi costs money and there are open source LLM's out there
-
I have followed the steps of the "How-To-Guide" about Automatic Language Adaptation (https://docs.ragas.io/en/v0.1.9/howtos/applications/use_prompt_adaptation.html) and tried to modify the already im…
-
having the ability to use the api to paid services is cute and all.
can we have local only.
nobody wants to pay for these services anymore especially as llama3.1 blew them away with costly tie…
-
Hello, I want to deploy llama-3-8b quantized model using tritonserver I followed below steps to do this:
1. create container with nvcr.io/nvidia/tritonserver:24.06-trtllm-python-py3 base image.
3.…
-
Wondering if possible to run with models like llama2 or huggingface models, or Ollama or something like litellm.
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [X] I am running the latest code. Development is very rapid so there are no tagged versions as of…
-
I am looking for mistral extractor for claim extraction process. The link seems to be broken.
-
## Description
I've been using the /generate command to create Jupyter notebooks from text prompts, but it seems to be generating filenames that contain colons (:). This is causing issues, espe…