Closed espritdunet closed 5 months ago
I have the same issue
After conducting further tests, I have realized that the Python 3.11 code in privateGPT does not seem to contain the necessary components for utilizing the local LM Studio server to leverage the "text embeddings" feature provided by it starting from version 0.2.19.
The settings-vllm.yaml model configuration I am using appears to lack the necessary details to connect to the local LM Studio server for using its new Text Embeddings function.
Here's some additional information:
I've consulted the LM Studio documentation available at: LM Studio Text Embeddings Documentation. My local tests for the connection work properly, as evident from the server logs of LM Studio and from my local prompts. For instance:
curl http://localhost:1234/v1/embeddings \
-H "Content-Type: application/json" \
-d '{
"input": "Your text string goes here",
"model": "model-identifier-here"
}'
However, I am uncertain about how to modify my privateGPT configuration file to accommodate this, and I am concerned that its current code may not support connecting in "openailike" mode to a Text Embeddings function supplied by LM Studio.
If time permits in the coming days, I will look into the privateGPT code to potentially integrate the connection to LM Studio for both the LLM and text embeddings models.
I appreciate any guidance on how to address this issue.
Thank you!
Hello everyone,
I wanted to provide an update on my issue regarding the integration of LM Studio's local server for LLM models and text embeddings with PrivateGPT.
Initially, I faced difficulties in making PrivateGPT communicate with the LM Studio server for the embedding model, as the current code of privateGPT did not support this setup. I found an alternative solution.
I utilized the huggingface_hub
Python library to download the text embedding models locally. This approach aligns with PrivateGPT's current functionalities and enabled me to implement the RAG (Retrieval-Augmented Generation) feature successfully.
Here are the steps I followed:
1. Install huggingface_hub
Library: I installed the huggingface_hub
library using pip:
pip install huggingface_hub
2. Login to Hugging Face: I logged in to my Hugging Face account using the CLI:
huggingface-cli login
I entered the token I created on the Hugging Face website when prompted.
poetry run python scripts/setup
After these steps, everything worked seamlessly, and I was able to run PrivateGPT with the desired setup. It turned out to be the standard procedure described in PrivateGPT's documentation, which I initially misunderstood.
I hope this update helps others who might be facing similar issues with LM Studio. If there are any further questions or if anyone needs more details on the implementation, feel free to reach out.
Thank you!
Hello,
I've installed privateGPT with Pyenv and Poetry on my MacBook M2 to set up a local RAG using LM Studio version 0.2.21.
I'm using the settings-vllm.yaml configuration file with the following setup:
I have successfully loaded the Nomic-AI and META models into LM Studio and it's working well. Its server is started on localhost:1234 and I may test it by running:
However, when I attempt to run privateGPT with the following commands:
I encounter the following error:
I'm puzzled why privateGPT is attempting to connect to OpenAI and Huggingface sites. What should I modify to correct this issue?
Any guidance would be greatly appreciated.
Thank you!