-
Notifications
You must be signed in to change notification settings - Fork 9.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feature Request: Support Jina V3 arch #9585
Comments
#6826 added the support for Jina V2. Perhaps something similar is needed. |
Jina V2 is based off of 'JinaBERT', V3 is based off of Jina-XLM-RoBERTa |
Okay, so the script already supports XLMRobertaModel. But the Flash implementation has different layer names or structures compared to the standard XLMRobertaModel, I believe. |
Maybe a fix similar to the one in #9510 would be needed: def modify_tensors(self, data_torch: Tensor, name: str, bid: int | None) -> Iterable[tuple[str, Tensor]]:
# if name starts with "roberta.", remove the prefix
# e.g. https://huggingface.co/BAAI/bge-reranker-v2-m3/tree/main
if name.startswith("roberta."):
name = name[8:] |
That might mitigate the immediate error, but later it will fall into another. For example, XLMRobertaModel looks for I made a very naive attempt of a whole lot of re-writing |
Prerequisites
Feature Description
I was trying to convert https://huggingface.co/jinaai/jina-embeddings-v3 to GGUF and it seems like it doesn't support it yet:
Motivation
Jina V3 has been one of the top performing embedding model. And it might be expected to see more models in the future.
Possible Implementation
No response
The text was updated successfully, but these errors were encountered: