Closed sigridjineth closed 1 month ago
TEI determines the type of backend model by the architectures
in config.json
code.
However, gte-multilingual-base
architectures in config.json
are NewModel
and NewForTokenClassification
. I guess it'd be good to change the config (e.g. removing NewForTokenClassification
from the architectures)
the reranker has a correct architecture name, but the config does not contain id2label
. So, I assume adding id2label
seems good (or loading as a reranker model if there's no id2label
key in the config.json).
In short, opening a PR to remove/add the config can solve the issues.
@nbroad1881 yup, I've found that renaming weight keys to remove prefixes like "new" and ensuring they follow the standard model naming pattern (e.g., "encoder.layer.0.attention.self.query.weight") improves compatibility for its very own.
but removing it would raise not using sparse weight predictions as mentioned so that need to look at more
Thanks for the PR on the repositories @nbroad1881! Yes, at the end of the day, model creators are free to follow whatever naming scheme, conventions... they want and they sometimes clash with the ones we setup in our repositories unfortunately.
Model description
The GTE model family has consistent support on Hugging Face TEI, but the newly introduced
gte-multilingual-base
fails to be supported at this moment.The same error happens for both biencoder and reranker.
Open source status
Provide useful links for the implementation
https://huggingface.co/Alibaba-NLP/gte-multilingual-base https://huggingface.co/Alibaba-NLP/gte-multilingual-reranker-base