Update README.md
Browse files
README.md
CHANGED
|
@@ -5,26 +5,5 @@ license: mit
|
|
| 5 |
|
| 6 |
## DeBERTa: Decoding-enhanced BERT with Disentangled Attention
|
| 7 |
|
| 8 |
-
[DeBERTa](https://
|
| 9 |
|
| 10 |
-
Please check the [official repository](https://github.com/microsoft/DeBERTa) for more details and updates.
|
| 11 |
-
|
| 12 |
-
This the DeBERTa V2 xlarge model fine-tuned with MNLI task, 24 layers, 1536 hidden size. Total parameters 900M.
|
| 13 |
-
|
| 14 |
-
## This model is deprecated, please use [DeBERTa-V2-XLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xlarge-mnli)
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
### Citation
|
| 18 |
-
|
| 19 |
-
If you find DeBERTa useful for your work, please cite the following paper:
|
| 20 |
-
|
| 21 |
-
``` latex
|
| 22 |
-
@inproceedings{
|
| 23 |
-
he2021deberta,
|
| 24 |
-
title={DEBERTA: DECODING-ENHANCED BERT WITH DISENTANGLED ATTENTION},
|
| 25 |
-
author={Pengcheng He and Xiaodong Liu and Jianfeng Gao and Weizhu Chen},
|
| 26 |
-
booktitle={International Conference on Learning Representations},
|
| 27 |
-
year={2021},
|
| 28 |
-
url={https://openreview.net/forum?id=XPZIaotutsD}
|
| 29 |
-
}
|
| 30 |
-
```
|
|
|
|
| 5 |
|
| 6 |
## DeBERTa: Decoding-enhanced BERT with Disentangled Attention
|
| 7 |
|
| 8 |
+
## This model is DEPRECATED, please use [DeBERTa-V2-XLarge-MNLI](https://huggingface.co/microsoft/deberta-v2-xlarge-mnli)
|
| 9 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|