Update README.md
Browse files
README.md
CHANGED
|
@@ -13,7 +13,37 @@ This model is trained through the approach described in [DMRetriever: A Family o
|
|
| 13 |
The associated GitHub repository is available [here](https://github.com/KaiYin97/DMRETRIEVER).
|
| 14 |
This model has 33M parameters.
|
| 15 |
|
| 16 |
-
##
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
|
| 18 |
Using HuggingFace Transformers:
|
| 19 |
```python
|
|
@@ -97,7 +127,7 @@ qa_embs = encode_texts(qa_queries)
|
|
| 97 |
print("QA Embeddings shape:", qa_embs.shape)
|
| 98 |
|
| 99 |
```
|
| 100 |
-
## Citation
|
| 101 |
If you find this repository helpful, please kindly consider citing the corresponding paper. Thanks!
|
| 102 |
```
|
| 103 |
@article{yin2025dmretriever,
|
|
|
|
| 13 |
The associated GitHub repository is available [here](https://github.com/KaiYin97/DMRETRIEVER).
|
| 14 |
This model has 33M parameters.
|
| 15 |
|
| 16 |
+
## 🧠 Model Overview
|
| 17 |
+
|
| 18 |
+
**DMRetriever-33M** has the following features:
|
| 19 |
+
|
| 20 |
+
- Model Type: Text Embedding
|
| 21 |
+
- Supported Languages: English
|
| 22 |
+
- Number of Paramaters: 33M
|
| 23 |
+
- Embedding Dimension: 384
|
| 24 |
+
|
| 25 |
+
For more details, including model training, benchmark evaluation, and inference performance, please refer to our [paper](https://www.arxiv.org/abs/2510.15087), [GitHub](https://github.com/KaiYin97/DMRETRIEVER).
|
| 26 |
+
|
| 27 |
+
|
| 28 |
+
## 📦 DMRetriever Series Model List
|
| 29 |
+
|
| 30 |
+
| **Model** | **Description** | **Backbone** | **Backbone Type** | **Hidden Size** | **#Layers** |
|
| 31 |
+
|:--|:--|:--|:--|:--:|:--:|
|
| 32 |
+
| [DMRetriever-33M](https://huggingface.co/DMIR01/DMRetriever-33M) | Base 33M variant | MiniLM | Encoder-only | 384 | 12 |
|
| 33 |
+
| [DMRetriever-33M-PT](https://huggingface.co/DMIR01/DMRetriever-33M-PT) | Pre-trained version of 33M | MiniLM | Encoder-only | 384 | 12 |
|
| 34 |
+
| [DMRetriever-109M](https://huggingface.co/DMIR01/DMRetriever-109M) | Base 109M variant | BERT-base-uncased | Encoder-only | 768 | 12 |
|
| 35 |
+
| [DMRetriever-109M-PT](https://huggingface.co/DMIR01/DMRetriever-109M-PT) | Pre-trained version of 109M | BERT-base-uncased | Encoder-only | 768 | 12 |
|
| 36 |
+
| [DMRetriever-335M](https://huggingface.co/DMIR01/DMRetriever-335M) | Base 335M variant | BERT-large-uncased-WWM | Encoder-only | 1024 | 24 |
|
| 37 |
+
| [DMRetriever-335M-PT](https://huggingface.co/DMIR01/DMRetriever-335M-PT) | Pre-trained version of 335M | BERT-large-uncased-WWM | Encoder-only | 1024 | 24 |
|
| 38 |
+
| [DMRetriever-596M](https://huggingface.co/DMIR01/DMRetriever-596M) | Base 596M variant | Qwen3-0.6B | Decoder-only | 1024 | 28 |
|
| 39 |
+
| [DMRetriever-596M-PT](https://huggingface.co/DMIR01/DMRetriever-596M-PT) | Pre-trained version of 596M | Qwen3-0.6B | Decoder-only | 1024 | 28 |
|
| 40 |
+
| [DMRetriever-4B](https://huggingface.co/DMIR01/DMRetriever-4B) | Base 4B variant | Qwen3-4B | Decoder-only | 2560 | 36 |
|
| 41 |
+
| [DMRetriever-4B-PT](https://huggingface.co/DMIR01/DMRetriever-4B-PT) | Pre-trained version of 4B | Qwen3-4B | Decoder-only | 2560 | 36 |
|
| 42 |
+
| [DMRetriever-7.6B](https://huggingface.co/DMIR01/DMRetriever-7.6B) | Base 7.6B variant | Qwen3-8B | Decoder-only | 4096 | 36 |
|
| 43 |
+
| [DMRetriever-7.6B-PT](https://huggingface.co/DMIR01/DMRetriever-7.6B-PT) | Pre-trained version of 7.6B | Qwen3-8B | Decoder-only | 4096 | 36 |
|
| 44 |
+
|
| 45 |
+
|
| 46 |
+
## 🚀 Usage
|
| 47 |
|
| 48 |
Using HuggingFace Transformers:
|
| 49 |
```python
|
|
|
|
| 127 |
print("QA Embeddings shape:", qa_embs.shape)
|
| 128 |
|
| 129 |
```
|
| 130 |
+
## 🧾 Citation
|
| 131 |
If you find this repository helpful, please kindly consider citing the corresponding paper. Thanks!
|
| 132 |
```
|
| 133 |
@article{yin2025dmretriever,
|