Commit
·
5cde4ed
1
Parent(s):
31a5652
Update README.md
Browse files
README.md
CHANGED
|
@@ -40,38 +40,53 @@ print("\n" + "-" * 50)
|
|
| 40 |
|
| 41 |
## Benchmarks
|
| 42 |
|
| 43 |
-
|
| 44 |
-
|
| 45 |
-
|
| 46 |
-
|
| 47 |
-
|[
|
| 48 |
-
|[
|
| 49 |
-
[
|
| 50 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 51 |
|
| 52 |
-
`*` - Weighted Average
|
| 53 |
|
| 54 |
-
The benchmarking datasets are as follows:
|
| 55 |
-
* **SC:** **[Sentiment Classification](https://ieeexplore.ieee.org/document/8554396/)**
|
| 56 |
-
* **EC:** **[Emotion Classification](https://aclanthology.org/2021.naacl-srw.19/)**
|
| 57 |
-
* **DC:** **[Document Classification](https://arxiv.org/abs/2005.00085)**
|
| 58 |
-
* **NER:** **[Named Entity Recognition](https://content.iospress.com/articles/journal-of-intelligent-and-fuzzy-systems/ifs179349)**
|
| 59 |
-
* **NLI:** **[Natural Language Inference](#datasets)**
|
| 60 |
|
| 61 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 62 |
## Citation
|
| 63 |
|
| 64 |
If you use this model, please cite the following paper:
|
| 65 |
```
|
| 66 |
-
@
|
| 67 |
-
|
| 68 |
-
|
| 69 |
-
|
| 70 |
-
|
| 71 |
-
|
| 72 |
-
|
| 73 |
-
|
| 74 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 75 |
}
|
| 76 |
```
|
| 77 |
|
|
|
|
| 40 |
|
| 41 |
## Benchmarks
|
| 42 |
|
| 43 |
+
* Zero-shot cross-lingual transfer-learning
|
| 44 |
+
|
| 45 |
+
| Model | Params | SC (macro-F1) | NLI (accuracy) | NER (micro-F1) | QA (EM/F1) | BangLUE score |
|
| 46 |
+
|----------------|-----------|-----------|-----------|-----------|-----------|-----------|
|
| 47 |
+
|[mBERT](https://huggingface.co/bert-base-multilingual-cased) | 180M | 27.05 | 62.22 | 39.27 | 59.01/64.18 | 50.35 |
|
| 48 |
+
|[XLM-R (base)](https://huggingface.co/xlm-roberta-base) | 270M | 42.03 | 72.18 | 45.37 | 55.03/61.83 | 55.29 |
|
| 49 |
+
|[XLM-R (large)](https://huggingface.co/xlm-roberta-large) | 550M | 68.96 | 78.16 | 57.74 | 71.13/77.70 | 70.74 |
|
| 50 |
+
|
| 51 |
+
* Supervised fine-tuning
|
| 52 |
+
|
| 53 |
+
| Model | Params | SC (macro-F1) | NLI (accuracy) | NER (micro-F1) | QA (EM/F1) | BangLUE score |
|
| 54 |
+
|----------------|-----------|-----------|-----------|-----------|-----------|-----------|
|
| 55 |
+
|[mBERT](https://huggingface.co/bert-base-multilingual-cased) | 180M | 67.59 | 75.13 | 68.97 | 67.12/72.64 | 70.29 |
|
| 56 |
+
|[XLM-R (base)](https://huggingface.co/xlm-roberta-base) | 270M | 69.54 | 78.46 | 73.32 | 68.09/74.27 | 72.82 |
|
| 57 |
+
|[XLM-R (large)](https://huggingface.co/xlm-roberta-large) | 550M | 70.97 | 82.40 | 78.39 | 73.15/79.06 | 76.79 |
|
| 58 |
+
|[sahajBERT](https://huggingface.co/neuropark/sahajBERT) | 18M | 71.12 | 76.92 | 70.94 | 65.48/70.69 | 71.03 |
|
| 59 |
+
|[BanglaBERT](https://huggingface.co/csebuetnlp/banglabert) | 110M | 72.89 | 82.80 | 77.78 | 72.63/79.34 | **77.09** |
|
| 60 |
|
|
|
|
| 61 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 62 |
|
| 63 |
|
| 64 |
+
The benchmarking datasets are as follows:
|
| 65 |
+
* **SC:** **[Sentiment Classification](https://aclanthology.org/2021.findings-emnlp.278)**
|
| 66 |
+
* **NER:** **[Named Entity Recognition](https://multiconer.github.io/competition)**
|
| 67 |
+
* **NLI:** **[Natural Language Inference](https://github.com/csebuetnlp/banglabert/#datasets)**
|
| 68 |
+
* **QA:** **[Question Answering](https://github.com/csebuetnlp/banglabert/#datasets)**
|
| 69 |
+
|
| 70 |
## Citation
|
| 71 |
|
| 72 |
If you use this model, please cite the following paper:
|
| 73 |
```
|
| 74 |
+
@inproceedings{bhattacharjee-etal-2022-banglabert,
|
| 75 |
+
title = {BanglaBERT: Lagnuage Model Pretraining and Benchmarks for Low-Resource Language Understanding Evaluation in Bangla},
|
| 76 |
+
author = "Bhattacharjee, Abhik and
|
| 77 |
+
Hasan, Tahmid and
|
| 78 |
+
Mubasshir, Kazi and
|
| 79 |
+
Islam, Md. Saiful and
|
| 80 |
+
Uddin, Wasi Ahmad and
|
| 81 |
+
Iqbal, Anindya and
|
| 82 |
+
Rahman, M. Sohel and
|
| 83 |
+
Shahriyar, Rifat",
|
| 84 |
+
booktitle = "Findings of the North American Chapter of the Association for Computational Linguistics: NAACL 2022",
|
| 85 |
+
month = july,
|
| 86 |
+
year = {2022},
|
| 87 |
+
url = {https://arxiv.org/abs/2101.00204},
|
| 88 |
+
eprinttype = {arXiv},
|
| 89 |
+
eprint = {2101.00204}
|
| 90 |
}
|
| 91 |
```
|
| 92 |
|