Update README.md
Browse files
README.md
CHANGED
|
@@ -102,53 +102,8 @@ Evaluation to come.
|
|
| 102 |
|
| 103 |
## Thank You and Credits
|
| 104 |
|
| 105 |
-
- To [occiglot](https://huggingface.co/occiglot) and OSCAR for their data used to pre-train the model
|
| 106 |
-
- To [deepset](https://huggingface.co/deepset) for the gbert-large, which is a really great model
|
| 107 |
- To [jinaAI](https://huggingface.co/jinaai) for their BERT implementation that is used, especially ALiBi
|
|
|
|
|
|
|
| 108 |
- To [Tom](https://huggingface.co/tomaarsen), especially for sentence-transformers, [Björn and Jan from ellamind](https://ellamind.com/de/) for the consultation
|
| 109 |
-
- To [Meta](https://huggingface.co/facebook) for XNLI
|
| 110 |
-
|
| 111 |
-
### BibTeX
|
| 112 |
-
|
| 113 |
-
#### Sentence Transformers
|
| 114 |
-
```bibtex
|
| 115 |
-
@inproceedings{reimers-2019-sentence-bert,
|
| 116 |
-
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
|
| 117 |
-
author = "Reimers, Nils and Gurevych, Iryna",
|
| 118 |
-
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
|
| 119 |
-
month = "11",
|
| 120 |
-
year = "2019",
|
| 121 |
-
publisher = "Association for Computational Linguistics",
|
| 122 |
-
url = "https://arxiv.org/abs/1908.10084",
|
| 123 |
-
}
|
| 124 |
-
```
|
| 125 |
-
|
| 126 |
-
#### MatryoshkaLoss
|
| 127 |
-
```bibtex
|
| 128 |
-
@misc{kusupati2024matryoshka,
|
| 129 |
-
title={Matryoshka Representation Learning},
|
| 130 |
-
author={Aditya Kusupati and Gantavya Bhatt and Aniket Rege and Matthew Wallingford and Aditya Sinha and Vivek Ramanujan and William Howard-Snyder and Kaifeng Chen and Sham Kakade and Prateek Jain and Ali Farhadi},
|
| 131 |
-
year={2024},
|
| 132 |
-
eprint={2205.13147},
|
| 133 |
-
archivePrefix={arXiv},
|
| 134 |
-
primaryClass={cs.LG}
|
| 135 |
-
}
|
| 136 |
-
```
|
| 137 |
-
|
| 138 |
-
<!--
|
| 139 |
-
## Glossary
|
| 140 |
-
|
| 141 |
-
*Clearly define terms in order to be accessible across audiences.*
|
| 142 |
-
-->
|
| 143 |
-
|
| 144 |
-
<!--
|
| 145 |
-
## Model Card Authors
|
| 146 |
-
|
| 147 |
-
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
|
| 148 |
-
-->
|
| 149 |
-
|
| 150 |
-
<!--
|
| 151 |
-
## Model Card Contact
|
| 152 |
-
|
| 153 |
-
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
|
| 154 |
-
-->
|
|
|
|
| 102 |
|
| 103 |
## Thank You and Credits
|
| 104 |
|
|
|
|
|
|
|
| 105 |
- To [jinaAI](https://huggingface.co/jinaai) for their BERT implementation that is used, especially ALiBi
|
| 106 |
+
- To [deepset](https://huggingface.co/deepset) for the gbert-large, which is a really great model
|
| 107 |
+
- To [occiglot](https://huggingface.co/occiglot) and OSCAR for their data used to pre-train the model
|
| 108 |
- To [Tom](https://huggingface.co/tomaarsen), especially for sentence-transformers, [Björn and Jan from ellamind](https://ellamind.com/de/) for the consultation
|
| 109 |
+
- To [Meta](https://huggingface.co/facebook) for XNLI which is used in variations
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|