Update model card with paper and GitHub links
#2
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,9 +1,10 @@
|
|
| 1 |
---
|
| 2 |
-
license: cc-by-nc-4.0
|
| 3 |
-
language:
|
| 4 |
-
- en
|
| 5 |
base_model:
|
| 6 |
- Qwen/Qwen3-4B
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
pipeline_tag: text-ranking
|
| 8 |
tags:
|
| 9 |
- finance
|
|
@@ -11,20 +12,23 @@ tags:
|
|
| 11 |
- code
|
| 12 |
- stem
|
| 13 |
- medical
|
| 14 |
-
library_name: sentence-transformers
|
| 15 |
---
|
| 16 |
|
| 17 |
<img src="https://i.imgur.com/oxvhvQu.png"/>
|
| 18 |
|
| 19 |
# Releasing zeroentropy/zerank-1
|
| 20 |
|
| 21 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 22 |
|
| 23 |
However, SOTA rerankers are closed-source and proprietary. At ZeroEntropy, we've trained a SOTA reranker outperforming closed-source competitors, and we're launching our model here on HuggingFace.
|
| 24 |
|
| 25 |
This reranker [outperforms proprietary rerankers](https://huggingface.co/zeroentropy/zerank-1#evaluations) such as `cohere-rerank-v3.5` and `Salesforce/LlamaRank-v1` across a wide variety of domains, including finance, legal, code, STEM, medical, and conversational data.
|
| 26 |
|
| 27 |
-
At ZeroEntropy we've developed an innovative multi-stage pipeline that models query-document relevance scores as adjusted [Elo ratings](https://en.wikipedia.org/wiki/Elo_rating_system).
|
| 28 |
|
| 29 |
Since we're a small company, this model is only released under a non-commercial license. If you'd like a commercial license, please contact us at [email protected] and we'll get you a license ASAP.
|
| 30 |
|
|
@@ -53,14 +57,14 @@ The model can also be inferenced using ZeroEntropy's [/models/rerank](https://do
|
|
| 53 |
|
| 54 |
NDCG@10 scores between `zerank-1` and competing closed-source proprietary rerankers. Since we are evaluating rerankers, OpenAI's `text-embedding-3-small` is used as an initial retriever for the Top 100 candidate documents.
|
| 55 |
|
| 56 |
-
| Task
|
| 57 |
-
|
| 58 |
-
| Code
|
| 59 |
-
| Conversational |
|
| 60 |
-
| Finance
|
| 61 |
-
| Legal
|
| 62 |
-
| Medical
|
| 63 |
-
| STEM
|
| 64 |
|
| 65 |
Comparing BM25 and Hybrid Search without and with zerank-1:
|
| 66 |
|
|
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
| 2 |
base_model:
|
| 3 |
- Qwen/Qwen3-4B
|
| 4 |
+
language:
|
| 5 |
+
- en
|
| 6 |
+
library_name: sentence-transformers
|
| 7 |
+
license: cc-by-nc-4.0
|
| 8 |
pipeline_tag: text-ranking
|
| 9 |
tags:
|
| 10 |
- finance
|
|
|
|
| 12 |
- code
|
| 13 |
- stem
|
| 14 |
- medical
|
|
|
|
| 15 |
---
|
| 16 |
|
| 17 |
<img src="https://i.imgur.com/oxvhvQu.png"/>
|
| 18 |
|
| 19 |
# Releasing zeroentropy/zerank-1
|
| 20 |
|
| 21 |
+
This model is the `zerank-1` reranker as introduced in the paper [zELO: ELO-inspired Training Method for Rerankers and Embedding Models](https://huggingface.co/papers/2509.12541).
|
| 22 |
+
|
| 23 |
+
**Code**: [https://github.com/zeroentropy-ai/zbench](https://github.com/zeroentropy-ai/zbench)
|
| 24 |
+
|
| 25 |
+
In search engines, [rerankers are crucial](https://www.zeroentropy.dev/blog/what-is-a-reranker-and-do_i_need_one) for improving the accuracy of your retrieval system.
|
| 26 |
|
| 27 |
However, SOTA rerankers are closed-source and proprietary. At ZeroEntropy, we've trained a SOTA reranker outperforming closed-source competitors, and we're launching our model here on HuggingFace.
|
| 28 |
|
| 29 |
This reranker [outperforms proprietary rerankers](https://huggingface.co/zeroentropy/zerank-1#evaluations) such as `cohere-rerank-v3.5` and `Salesforce/LlamaRank-v1` across a wide variety of domains, including finance, legal, code, STEM, medical, and conversational data.
|
| 30 |
|
| 31 |
+
At ZeroEntropy we've developed an innovative multi-stage pipeline that models query-document relevance scores as adjusted [Elo ratings](https://en.wikipedia.org/wiki/Elo_rating_system). More details are available in our paper: [zELO: ELO-inspired Training Method for Rerankers and Embedding Models](https://huggingface.co/papers/2509.12541).
|
| 32 |
|
| 33 |
Since we're a small company, this model is only released under a non-commercial license. If you'd like a commercial license, please contact us at [email protected] and we'll get you a license ASAP.
|
| 34 |
|
|
|
|
| 57 |
|
| 58 |
NDCG@10 scores between `zerank-1` and competing closed-source proprietary rerankers. Since we are evaluating rerankers, OpenAI's `text-embedding-3-small` is used as an initial retriever for the Top 100 candidate documents.
|
| 59 |
|
| 60 |
+
| Task | Embedding | cohere-rerank-v3.5 | Salesforce/Llama-rank-v1 | zerank-1-small | **zerank-1** |
|
| 61 |
+
|---|---|---|---|---|---|
|
| 62 |
+
| Code | 0.678 | 0.724 | 0.694 | 0.730 | **0.754** |
|
| 63 |
+
| Conversational | 0.250 | 0.571 | 0.484 | 0.556 | **0.596** |
|
| 64 |
+
| Finance | 0.839 | 0.824 | 0.828 | 0.861 | **0.894** |
|
| 65 |
+
| Legal | 0.703 | 0.804 | 0.767 | 0.817 | **0.821** |
|
| 66 |
+
| Medical | 0.619 | 0.750 | 0.719 | 0.773 | **0.796** |
|
| 67 |
+
| STEM | 0.401 | 0.510 | 0.595 | 0.680 | **0.694** |
|
| 68 |
|
| 69 |
Comparing BM25 and Hybrid Search without and with zerank-1:
|
| 70 |
|