Add library_name and pipeline_tag metadata (#1)
Browse files- Add library_name and pipeline_tag metadata (fe072c4fe8ad0375cc4bf07f9af1b1172c911798)
Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>
README.md
CHANGED
|
@@ -1,19 +1,24 @@
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
license: other
|
| 3 |
license_name: rail-share
|
| 4 |
license_link: LICENSE
|
| 5 |
-
language:
|
| 6 |
-
- en
|
| 7 |
metrics:
|
| 8 |
- perplexity
|
| 9 |
-
|
| 10 |
-
|
| 11 |
---
|
| 12 |
|
| 13 |
# Model Card for SHARE-14B
|
| 14 |
|
| 15 |
SHARE-14B (Social-Humanities AI for Research and Education) is a 14-billion-parameter decoder-only causal language model pretrained exclusively on content relevant to the social sciences and humanities (SSH). It is intended as a domain-specific base model for SSH research and education, and is designed to be used through the MIRROR interface, which surfaces token-level surprisal rather than generating new text.
|
| 16 |
|
|
|
|
|
|
|
| 17 |
**Note:** This is an intermediate checkpoint released after ~15% of planned pretraining (96B tokens of a target ~630B). It is a base (pretrained-only) model with no SFT, DPO, or RLHF. This base model is not suitable to chat applications.
|
| 18 |
|
| 19 |
## Model Details
|
|
@@ -31,8 +36,7 @@ SHARE-14B is the first causal language model fully pretrained by and for the SSH
|
|
| 31 |
### Model Sources
|
| 32 |
|
| 33 |
- **Repository:** https://github.com/Joaoffg/SHARE
|
| 34 |
-
- **Paper:** SHARE
|
| 35 |
-
- **Demo (MIRROR interface):** [Add link]
|
| 36 |
- **Contact:** ferreiragoncalves@eshcc.eur.nl
|
| 37 |
|
| 38 |
## Uses
|
|
|
|
| 1 |
---
|
| 2 |
+
datasets:
|
| 3 |
+
- allenai/peS2o
|
| 4 |
+
language:
|
| 5 |
+
- en
|
| 6 |
+
- nl
|
| 7 |
license: other
|
| 8 |
license_name: rail-share
|
| 9 |
license_link: LICENSE
|
|
|
|
|
|
|
| 10 |
metrics:
|
| 11 |
- perplexity
|
| 12 |
+
library_name: transformers
|
| 13 |
+
pipeline_tag: text-generation
|
| 14 |
---
|
| 15 |
|
| 16 |
# Model Card for SHARE-14B
|
| 17 |
|
| 18 |
SHARE-14B (Social-Humanities AI for Research and Education) is a 14-billion-parameter decoder-only causal language model pretrained exclusively on content relevant to the social sciences and humanities (SSH). It is intended as a domain-specific base model for SSH research and education, and is designed to be used through the MIRROR interface, which surfaces token-level surprisal rather than generating new text.
|
| 19 |
|
| 20 |
+
More information can be found in the paper [SHARE: Social-Humanities AI for Research and Education](https://huggingface.co/papers/2604.11152).
|
| 21 |
+
|
| 22 |
**Note:** This is an intermediate checkpoint released after ~15% of planned pretraining (96B tokens of a target ~630B). It is a base (pretrained-only) model with no SFT, DPO, or RLHF. This base model is not suitable to chat applications.
|
| 23 |
|
| 24 |
## Model Details
|
|
|
|
| 36 |
### Model Sources
|
| 37 |
|
| 38 |
- **Repository:** https://github.com/Joaoffg/SHARE
|
| 39 |
+
- **Paper:** [SHARE: Social-Humanities AI for Research and Education](https://arxiv.org/abs/2604.11152)
|
|
|
|
| 40 |
- **Contact:** ferreiragoncalves@eshcc.eur.nl
|
| 41 |
|
| 42 |
## Uses
|