-
Hi,
I'm new to Langchain and LLM.
I've recently deployed an LLM model using the Hugging Face text-generation-inference library on my local machine.
I've successfully accessed the model using …
-
### 🥰 Feature Description
Right Now in interface setting there is one option to set openai model after setting it we cant create another openai custom model. I am using LiteLLM so i need more than on…
-
Instead of using chat-gpt, I would like to try and use a local LLM. I am sure this would take some modifications, but I think we could potentially make this work, and would be an awesome addition to t…
-
### **Title**: Dynamic FAISS Integration for RAG in Prototyping Conversations
---
### **Description**
Implement a lightweight Retrieval-Augmented Generation (RAG) solution using FAISS for small-sc…
-
### Current Behavior
When I implement GPTCache according to Documentation it does not work.
I am using the GPTCache adapter of langchain and the Langchain Adapter for my embedding
In the end i …
-
Getting the below error when trying to run the Llam2 70B benchmark as given in the link - [Here](https://github.com/HabanaAI/Model-References/tree/master/MLPERF4.0/Training/benchmarks/llm_finetune) wi…
-
Our LLM model is on endpoint :
# from langchain.indexes import VectorstoreIndexCreator
os.environ["OPENAI_API_BASE"] = "my_endpoint_ip"
os.environ["OPENAI_API_KEY"] = "empty"
embedding = OpenAIEmb…
-
Hi,
I have finetuned Qwen2-VL using Llama-Factory.
I successfully quantized the fine-tuned model as given
```
from transformers import Qwen2VLProcessor
from auto_gptq import BaseQuantizeC…
-
Search is NOT limited to given txt file.
`
from crewai_tools import TXTSearchTool
txt_search_tool = TXTSearchTool(
txt="kunst.txt",
config=dict(
llm=dict(
provid…
-
I try to use preble to deploy a model by sglang, but get an error:
```
$ preble run -port 6666 -model /workspace/LLMs/Qwen2-7B-Instruct
Traceback (most recent call last):
File "/usr/local/bin/…