KaiYinTAMU commited on
Commit
0940fcb
·
verified ·
1 Parent(s): 6af1306

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +31 -2
README.md CHANGED
@@ -13,7 +13,36 @@ This model is trained through the approach described in [DMRetriever: A Family o
13
  The associated GitHub repository is available [here](https://github.com/KaiYin97/DMRETRIEVER).
14
  This model has 335M parameters and it is the pre-trained version (trained using only unlabeled dataset containing in-batch negative).
15
 
16
- ## Usage
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
 
18
  Using HuggingFace Transformers:
19
  ```python
@@ -97,7 +126,7 @@ qa_embs = encode_texts(qa_queries)
97
  print("QA Embeddings shape:", qa_embs.shape)
98
 
99
  ```
100
- ## Citation
101
  If you find this repository helpful, please kindly consider citing the corresponding paper. Thanks!
102
  ```
103
  @article{yin2025dmretriever,
 
13
  The associated GitHub repository is available [here](https://github.com/KaiYin97/DMRETRIEVER).
14
  This model has 335M parameters and it is the pre-trained version (trained using only unlabeled dataset containing in-batch negative).
15
 
16
+ ## 🧠 Model Overview
17
+
18
+ **DMRetriever-335M-PT** has the following features:
19
+
20
+ - Model Type: Text Embedding
21
+ - Supported Languages: English
22
+ - Number of Paramaters: 335M
23
+ - Embedding Dimension: 1024
24
+
25
+ For more details, including model training, benchmark evaluation, and inference performance, please refer to our [paper](https://www.arxiv.org/abs/2510.15087), [GitHub](https://github.com/KaiYin97/DMRETRIEVER).
26
+
27
+ ## 📦 DMRetriever Series Model List
28
+
29
+ | **Model** | **Description** | **Backbone** | **Backbone Type** | **Hidden Size** | **#Layers** |
30
+ |:--|:--|:--|:--|:--:|:--:|
31
+ | [DMRetriever-33M](https://huggingface.co/DMIR01/DMRetriever-33M) | Base 33M variant | MiniLM | Encoder-only | 384 | 12 |
32
+ | [DMRetriever-33M-PT](https://huggingface.co/DMIR01/DMRetriever-33M-PT) | Pre-trained version of 33M | MiniLM | Encoder-only | 384 | 12 |
33
+ | [DMRetriever-109M](https://huggingface.co/DMIR01/DMRetriever-109M) | Base 109M variant | BERT-base-uncased | Encoder-only | 768 | 12 |
34
+ | [DMRetriever-109M-PT](https://huggingface.co/DMIR01/DMRetriever-109M-PT) | Pre-trained version of 109M | BERT-base-uncased | Encoder-only | 768 | 12 |
35
+ | [DMRetriever-335M](https://huggingface.co/DMIR01/DMRetriever-335M) | Base 335M variant | BERT-large-uncased-WWM | Encoder-only | 1024 | 24 |
36
+ | [DMRetriever-335M-PT](https://huggingface.co/DMIR01/DMRetriever-335M-PT) | Pre-trained version of 335M | BERT-large-uncased-WWM | Encoder-only | 1024 | 24 |
37
+ | [DMRetriever-596M](https://huggingface.co/DMIR01/DMRetriever-596M) | Base 596M variant | Qwen3-0.6B | Decoder-only | 1024 | 28 |
38
+ | [DMRetriever-596M-PT](https://huggingface.co/DMIR01/DMRetriever-596M-PT) | Pre-trained version of 596M | Qwen3-0.6B | Decoder-only | 1024 | 28 |
39
+ | [DMRetriever-4B](https://huggingface.co/DMIR01/DMRetriever-4B) | Base 4B variant | Qwen3-4B | Decoder-only | 2560 | 36 |
40
+ | [DMRetriever-4B-PT](https://huggingface.co/DMIR01/DMRetriever-4B-PT) | Pre-trained version of 4B | Qwen3-4B | Decoder-only | 2560 | 36 |
41
+ | [DMRetriever-7.6B](https://huggingface.co/DMIR01/DMRetriever-7.6B) | Base 7.6B variant | Qwen3-8B | Decoder-only | 4096 | 36 |
42
+ | [DMRetriever-7.6B-PT](https://huggingface.co/DMIR01/DMRetriever-7.6B-PT) | Pre-trained version of 7.6B | Qwen3-8B | Decoder-only | 4096 | 36 |
43
+
44
+
45
+ ## 🚀 Usage
46
 
47
  Using HuggingFace Transformers:
48
  ```python
 
126
  print("QA Embeddings shape:", qa_embs.shape)
127
 
128
  ```
129
+ ## 🧾 Citation
130
  If you find this repository helpful, please kindly consider citing the corresponding paper. Thanks!
131
  ```
132
  @article{yin2025dmretriever,