Vít Novotný
commited on
Commit
·
12b87ed
1
Parent(s):
204c0b9
Document huggingface/transformers#16936 in `README.md`
Browse files
README.md
CHANGED
|
@@ -6,7 +6,7 @@ datasets:
|
|
| 6 |
- math-stackexchange
|
| 7 |
---
|
| 8 |
|
| 9 |
-
# MathBERTa
|
| 10 |
|
| 11 |
Pretrained model on English language and LaTeX using a masked language modeling
|
| 12 |
(MLM) objective. It was developed for [the ARQMath-3 shared task evaluation][1]
|
|
@@ -48,6 +48,11 @@ text generation you should look at model like GPT2.
|
|
| 48 |
|
| 49 |
### How to use
|
| 50 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 51 |
You can use this model directly with a pipeline for masked language modeling:
|
| 52 |
|
| 53 |
```python
|
|
@@ -99,3 +104,5 @@ Together theses datasets weight 52GB of text and LaTeX.
|
|
| 99 |
|
| 100 |
[5]: https://sigmathling.kwarc.info/resources/arxmliv-dataset-2020/
|
| 101 |
[6]: https://www.cs.rit.edu/~dprl/ARQMath/arqmath-resources.html
|
|
|
|
|
|
|
|
|
| 6 |
- math-stackexchange
|
| 7 |
---
|
| 8 |
|
| 9 |
+
# MathBERTa model
|
| 10 |
|
| 11 |
Pretrained model on English language and LaTeX using a masked language modeling
|
| 12 |
(MLM) objective. It was developed for [the ARQMath-3 shared task evaluation][1]
|
|
|
|
| 48 |
|
| 49 |
### How to use
|
| 50 |
|
| 51 |
+
|
| 52 |
+
*Due to the large number of added LaTeX tokens, MathBERTa is affected by [a
|
| 53 |
+
software bug in the 🤗 Transformers library][9] that causes it to load for tens
|
| 54 |
+
of minutes. The bug is [to be fixed in 🤗 Transformers 4.20.0][10].*
|
| 55 |
+
|
| 56 |
You can use this model directly with a pipeline for masked language modeling:
|
| 57 |
|
| 58 |
```python
|
|
|
|
| 104 |
|
| 105 |
[5]: https://sigmathling.kwarc.info/resources/arxmliv-dataset-2020/
|
| 106 |
[6]: https://www.cs.rit.edu/~dprl/ARQMath/arqmath-resources.html
|
| 107 |
+
[9]: https://github.com/huggingface/transformers/issues/16936
|
| 108 |
+
[10]: https://github.com/huggingface/transformers/pull/17119
|