FinBERT2-base / README.md
mengyouHF's picture
Update README.md
804a488 verified
|
raw
history blame
1.22 kB
---
license: apache-2.0
language:
- zh
base_model:
- hfl/chinese-roberta-wwm-ext
tags:
- finance
---
## Model Details
**Model Description:** This is a finance-domain pretrained Chinese language model, which is based on the 125-million-parameter RoBERTa-Base and further pre-trained on 32B tokens of Chinese financial corpora (including a large number of research reports, news, and announcements).
- **Developed by:** See [valuesimplex](https://github.com/valuesimplex) for model developers
- **Model Type:** Transformer-based language model
- **Language(s):** Chinese
- **Parent Model:** See the [chinese-roberta](https://huggingface.co/hfl/chinese-roberta-wwm-ext) for more information about the BERT base model.
- **Resources for more information:**
- [Research Paper](https://dl.acm.org/doi/10.1145/3711896.3737219)
- [GitHub Repo](https://github.com/valuesimplex/FinBERT)
## Direct Use
```python
from transformers import AutoModel, AutoTokenizer
model = AutoModel.from_pretrained("valuesimplex-ai-lab/FinBERT2-base")
tokenizer = AutoTokenizer.from_pretrained("valuesimplex-ai-lab/FinBERT2-base")
```
### Further Usage
continual pre-training or fine-tuning:https://github.com/valuesimplex/FinBERT