Open ruifengma opened 7 months ago
I managed to use GritLM as an embedding model, but the values are not right. Maybe you have more success than I do...
I managed to use GritLM as an embedding model, but the values are not right. Maybe you have more success than I do...
Hi @surak , Thanks for the reply. May I ask how do you deploy your embedding model?
I managed to use GritLM as an embedding model, but the values are not right. Maybe you have more success than I do... Hi @surak , Thanks for the reply. May I ask how do you deploy your embedding model?
Using the normal model_worker - the sglang and the vllm workers don't work.
I run it on slurm, but it's the same as anywhere else:
srun python3 $BLABLADOR_DIR/fastchat/serve/model_worker.py \
--controller $BLABLADOR_CONTROLLER:$BLABLADOR_CONTROLLER_PORT \
--port 31041 --worker-address http://$(hostname).fz-juelich.de:31041 \
--num-gpus 1 \
--host $BLABLADOR_CONTROLLER \
--model-path models/GritLM-7B \
--model-name "alias-embeddings,gpt-3.5-turbo,text-davinci-003,text-embedding-ada-002,GritLM-7B"
I alias to the openai model names so langchain can work.
I tried to use model_work to deploy an Embedding model gte-base
and I got the error which is
I use the transformers and pytorch with the code on the repo model card in the same environment and it works normally
So I would like to ask if FastChat support this embedding or model_worker is the correct to deploy embedding model? Thanks in advance