Closed majestichou closed 5 months ago
@codingl2k1 can you give a help?
@codingl2k1 can you give a help?
I am looking into this issue.
This model works well on my Mac, and I can run the rerank benchmark on it, though it's very slow.
Name: FlagEmbedding Version: 1.2.8
Name: transformers Version: 4.39.1
This model works well on my Mac, and I can run the rerank benchmark on it, though it's very slow.
Name: FlagEmbedding Version: 1.2.8
Name: transformers Version: 4.39.1
Which version of Xinference docker image did you use?
This model works well on my Mac, and I can run the rerank benchmark on it, though it's very slow.
Name: FlagEmbedding Version: 1.2.8
Name: transformers Version: 4.39.1
Can you try to repeat my steps below?
I downloaded the "bge-reranker-v2-minicpm-layerwise" model weights to the server and registered this model (the registered model name is "bge-reranker-v2-minicpm-layerwise-self") with Xinference v0.11.1 version (docker images). Then I launched this model. However, it crushed. The error information is as follows:
Make sure you have read the code there to avoid malicious use, then set the option
trust_remote_code=Trueto remove this error.
This model works well on my Mac, and I can run the rerank benchmark on it, though it's very slow. Name: FlagEmbedding Version: 1.2.8 Name: transformers Version: 4.39.1
Can you try to repeat my steps below? I downloaded the "bge-reranker-v2-minicpm-layerwise" model weights to the server and registered this model (the registered model name is "bge-reranker-v2-minicpm-layerwise-self") with Xinference v0.11.1 version (docker images). Then I launched this model. However, it crushed. The error information is as follows:
Make sure you have read the code there to avoid malicious use, then set the option
trust_remote_code=Trueto remove this error.
I will try your steps.
How do you register the model bge-reranker-v2-minicpm-layerwise-self
? The model spec's type should be LLM-based layerwise
. From your traceback, the callstack was in type == "normal"
.
How do you register the model
bge-reranker-v2-minicpm-layerwise-self
? The model spec's type should beLLM-based layerwise
. From your traceback, the callstack was intype == "normal"
.
I choose Register Model Tab,select "RERANK MODEL",fill in the parameters: model name and model path in docker container and choose English and Chinese. Finally,register.
Describe the bug
I downloaded the "bge-reranker-v2-minicpm-layerwise" model weights to the server and registered this model (the registered model name is "bge-reranker-v2-minicpm-layerwise-self") with Xinference v0.11.0 version (docker images). Then I launched this model. However, it crushed. The error information is as follows:
To Reproduce
Expected behavior
No crush. According the Xinference docs (https://inference.readthedocs.io/en/latest/models/builtin/rerank/bge-reranker-v2-minicpm-layerwise.html),
bge-reranker-v2-minicpm-layerwise model
is supported.