Closed abhishekbhakat closed 4 days ago
Jina V2 is based off of 'JinaBERT', V3 is based off of Jina-XLM-RoBERTa
Okay, so the script already supports XLMRobertaModel. But the Flash implementation has different layer names or structures compared to the standard XLMRobertaModel, I believe.
Maybe a fix similar to the one in #9510 would be needed:
def modify_tensors(self, data_torch: Tensor, name: str, bid: int | None) -> Iterable[tuple[str, Tensor]]:
# if name starts with "roberta.", remove the prefix
# e.g. https://huggingface.co/BAAI/bge-reranker-v2-m3/tree/main
if name.startswith("roberta."):
name = name[8:]
That might mitigate the immediate error, but later it will fall into another.
For example, XLMRobertaModel looks for sentencepiece.bpe.model
in the hf repo. But this repo only has a tokenizer.json
.
I made a very naive attempt of a whole lot of re-writing set_vocab()
but in the end got stuck in GGUFWriter getting a boolean in the array expecting <GGUFValueType.STRING: 8>
type 🥲.
This issue was closed because it has been inactive for 14 days since being marked as stale.
Prerequisites
Feature Description
I was trying to convert https://huggingface.co/jinaai/jina-embeddings-v3 to GGUF and it seems like it doesn't support it yet:
Motivation
Jina V3 has been one of the top performing embedding model. And it might be expected to see more models in the future.
Possible Implementation
No response