Update model card with paper and GitHub links

#2
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +18 -14
README.md CHANGED
@@ -1,9 +1,10 @@
1
  ---
2
- license: cc-by-nc-4.0
3
- language:
4
- - en
5
  base_model:
6
  - Qwen/Qwen3-4B
 
 
 
 
7
  pipeline_tag: text-ranking
8
  tags:
9
  - finance
@@ -11,20 +12,23 @@ tags:
11
  - code
12
  - stem
13
  - medical
14
- library_name: sentence-transformers
15
  ---
16
 
17
  <img src="https://i.imgur.com/oxvhvQu.png"/>
18
 
19
  # Releasing zeroentropy/zerank-1
20
 
21
- In search enginers, [rerankers are crucial](https://www.zeroentropy.dev/blog/what-is-a-reranker-and-do-i-need-one) for improving the accuracy of your retrieval system.
 
 
 
 
22
 
23
  However, SOTA rerankers are closed-source and proprietary. At ZeroEntropy, we've trained a SOTA reranker outperforming closed-source competitors, and we're launching our model here on HuggingFace.
24
 
25
  This reranker [outperforms proprietary rerankers](https://huggingface.co/zeroentropy/zerank-1#evaluations) such as `cohere-rerank-v3.5` and `Salesforce/LlamaRank-v1` across a wide variety of domains, including finance, legal, code, STEM, medical, and conversational data.
26
 
27
- At ZeroEntropy we've developed an innovative multi-stage pipeline that models query-document relevance scores as adjusted [Elo ratings](https://en.wikipedia.org/wiki/Elo_rating_system). See our Technical Report (Coming soon!) for more details.
28
 
29
  Since we're a small company, this model is only released under a non-commercial license. If you'd like a commercial license, please contact us at [email protected] and we'll get you a license ASAP.
30
 
@@ -53,14 +57,14 @@ The model can also be inferenced using ZeroEntropy's [/models/rerank](https://do
53
 
54
  NDCG@10 scores between `zerank-1` and competing closed-source proprietary rerankers. Since we are evaluating rerankers, OpenAI's `text-embedding-3-small` is used as an initial retriever for the Top 100 candidate documents.
55
 
56
- | Task | Embedding | cohere-rerank-v3.5 | Salesforce/Llama-rank-v1 | zerank-1-small | **zerank-1** |
57
- |----------------|-----------|--------------------|--------------------------|----------------|--------------|
58
- | Code | 0.678 | 0.724 | 0.694 | 0.730 | **0.754** |
59
- | Conversational | 0.250 | 0.571 | 0.484 | 0.556 | **0.596** |
60
- | Finance | 0.839 | 0.824 | 0.828 | 0.861 | **0.894** |
61
- | Legal | 0.703 | 0.804 | 0.767 | 0.817 | **0.821** |
62
- | Medical | 0.619 | 0.750 | 0.719 | 0.773 | **0.796** |
63
- | STEM | 0.401 | 0.510 | 0.595 | 0.680 | **0.694** |
64
 
65
  Comparing BM25 and Hybrid Search without and with zerank-1:
66
 
 
1
  ---
 
 
 
2
  base_model:
3
  - Qwen/Qwen3-4B
4
+ language:
5
+ - en
6
+ library_name: sentence-transformers
7
+ license: cc-by-nc-4.0
8
  pipeline_tag: text-ranking
9
  tags:
10
  - finance
 
12
  - code
13
  - stem
14
  - medical
 
15
  ---
16
 
17
  <img src="https://i.imgur.com/oxvhvQu.png"/>
18
 
19
  # Releasing zeroentropy/zerank-1
20
 
21
+ This model is the `zerank-1` reranker as introduced in the paper [zELO: ELO-inspired Training Method for Rerankers and Embedding Models](https://huggingface.co/papers/2509.12541).
22
+
23
+ **Code**: [https://github.com/zeroentropy-ai/zbench](https://github.com/zeroentropy-ai/zbench)
24
+
25
+ In search engines, [rerankers are crucial](https://www.zeroentropy.dev/blog/what-is-a-reranker-and-do_i_need_one) for improving the accuracy of your retrieval system.
26
 
27
  However, SOTA rerankers are closed-source and proprietary. At ZeroEntropy, we've trained a SOTA reranker outperforming closed-source competitors, and we're launching our model here on HuggingFace.
28
 
29
  This reranker [outperforms proprietary rerankers](https://huggingface.co/zeroentropy/zerank-1#evaluations) such as `cohere-rerank-v3.5` and `Salesforce/LlamaRank-v1` across a wide variety of domains, including finance, legal, code, STEM, medical, and conversational data.
30
 
31
+ At ZeroEntropy we've developed an innovative multi-stage pipeline that models query-document relevance scores as adjusted [Elo ratings](https://en.wikipedia.org/wiki/Elo_rating_system). More details are available in our paper: [zELO: ELO-inspired Training Method for Rerankers and Embedding Models](https://huggingface.co/papers/2509.12541).
32
 
33
  Since we're a small company, this model is only released under a non-commercial license. If you'd like a commercial license, please contact us at [email protected] and we'll get you a license ASAP.
34
 
 
57
 
58
  NDCG@10 scores between `zerank-1` and competing closed-source proprietary rerankers. Since we are evaluating rerankers, OpenAI's `text-embedding-3-small` is used as an initial retriever for the Top 100 candidate documents.
59
 
60
+ | Task | Embedding | cohere-rerank-v3.5 | Salesforce/Llama-rank-v1 | zerank-1-small | **zerank-1** |
61
+ |---|---|---|---|---|---|
62
+ | Code | 0.678 | 0.724 | 0.694 | 0.730 | **0.754** |
63
+ | Conversational | 0.250 | 0.571 | 0.484 | 0.556 | **0.596** |
64
+ | Finance | 0.839 | 0.824 | 0.828 | 0.861 | **0.894** |
65
+ | Legal | 0.703 | 0.804 | 0.767 | 0.817 | **0.821** |
66
+ | Medical | 0.619 | 0.750 | 0.719 | 0.773 | **0.796** |
67
+ | STEM | 0.401 | 0.510 | 0.595 | 0.680 | **0.694** |
68
 
69
  Comparing BM25 and Hybrid Search without and with zerank-1:
70