Update README.md
Browse files
README.md
CHANGED
|
@@ -8,13 +8,11 @@ tags:
|
|
| 8 |
- MarIA
|
| 9 |
---
|
| 10 |
# PlanTL-GOB-ES-roberta-base-bne
|
| 11 |
-
https://huggingface.co/PlanTL-GOB-ES/roberta-base-bne
|
| 12 |
-
|
| 13 |
-
**Copy of** MarIA ( **PlanTL-GOB-ES/roberta-base-bne** ) since weight files for this model has been removed permanently because it has been deprecated.
|
| 14 |
-
|
| 15 |
|
| 16 |
©© **All rights reserved: https://huggingface.co/PlanTL-GOB-ES/roberta-base-bne** ©©
|
| 17 |
|
|
|
|
|
|
|
| 18 |
|
| 19 |
## How to use
|
| 20 |
Here is how to use this model:
|
|
@@ -51,7 +49,7 @@ Here is how to use this model to get the features of a given text in PyTorch:
|
|
| 51 |
```python
|
| 52 |
>>> from transformers import RobertaTokenizer, RobertaModel
|
| 53 |
>>> tokenizer = RobertaTokenizer.from_pretrained('PeterPanecillo/PlanTL-GOB-ES-roberta-base-bne-copy')
|
| 54 |
-
>>> model = RobertaModel.from_pretrained('PlanTL-GOB-ES
|
| 55 |
>>> text = "Gracias a los datos de la BNE se ha podido desarrollar este modelo del lenguaje."
|
| 56 |
>>> encoded_input = tokenizer(text, return_tensors='pt')
|
| 57 |
>>> output = model(**encoded_input)
|
|
|
|
| 8 |
- MarIA
|
| 9 |
---
|
| 10 |
# PlanTL-GOB-ES-roberta-base-bne
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
|
| 12 |
©© **All rights reserved: https://huggingface.co/PlanTL-GOB-ES/roberta-base-bne** ©©
|
| 13 |
|
| 14 |
+
**Copy of** MarIA ( **PlanTL-GOB-ES/roberta-base-bne** ) since weight files for this model has been removed permanently because it has been deprecated.
|
| 15 |
+
|
| 16 |
|
| 17 |
## How to use
|
| 18 |
Here is how to use this model:
|
|
|
|
| 49 |
```python
|
| 50 |
>>> from transformers import RobertaTokenizer, RobertaModel
|
| 51 |
>>> tokenizer = RobertaTokenizer.from_pretrained('PeterPanecillo/PlanTL-GOB-ES-roberta-base-bne-copy')
|
| 52 |
+
>>> model = RobertaModel.from_pretrained('PeterPanecillo/PlanTL-GOB-ES-roberta-base-bne-copy')
|
| 53 |
>>> text = "Gracias a los datos de la BNE se ha podido desarrollar este modelo del lenguaje."
|
| 54 |
>>> encoded_input = tokenizer(text, return_tensors='pt')
|
| 55 |
>>> output = model(**encoded_input)
|