Joaoffg nielsr HF Staff commited on
Commit
728b25f
·
1 Parent(s): d1cb830

Add library_name and pipeline_tag metadata (#1)

Browse files

- Add library_name and pipeline_tag metadata (fe072c4fe8ad0375cc4bf07f9af1b1172c911798)


Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +10 -6
README.md CHANGED
@@ -1,19 +1,24 @@
1
  ---
 
 
 
 
 
2
  license: other
3
  license_name: rail-share
4
  license_link: LICENSE
5
- language:
6
- - en
7
  metrics:
8
  - perplexity
9
- datasets:
10
- - allenai/peS2o
11
  ---
12
 
13
  # Model Card for SHARE-14B
14
 
15
  SHARE-14B (Social-Humanities AI for Research and Education) is a 14-billion-parameter decoder-only causal language model pretrained exclusively on content relevant to the social sciences and humanities (SSH). It is intended as a domain-specific base model for SSH research and education, and is designed to be used through the MIRROR interface, which surfaces token-level surprisal rather than generating new text.
16
 
 
 
17
  **Note:** This is an intermediate checkpoint released after ~15% of planned pretraining (96B tokens of a target ~630B). It is a base (pretrained-only) model with no SFT, DPO, or RLHF. This base model is not suitable to chat applications.
18
 
19
  ## Model Details
@@ -31,8 +36,7 @@ SHARE-14B is the first causal language model fully pretrained by and for the SSH
31
  ### Model Sources
32
 
33
  - **Repository:** https://github.com/Joaoffg/SHARE
34
- - **Paper:** SHARE Technical Report: Social-Humanities AI for Research and Education (2026)
35
- - **Demo (MIRROR interface):** [Add link]
36
  - **Contact:** ferreiragoncalves@eshcc.eur.nl
37
 
38
  ## Uses
 
1
  ---
2
+ datasets:
3
+ - allenai/peS2o
4
+ language:
5
+ - en
6
+ - nl
7
  license: other
8
  license_name: rail-share
9
  license_link: LICENSE
 
 
10
  metrics:
11
  - perplexity
12
+ library_name: transformers
13
+ pipeline_tag: text-generation
14
  ---
15
 
16
  # Model Card for SHARE-14B
17
 
18
  SHARE-14B (Social-Humanities AI for Research and Education) is a 14-billion-parameter decoder-only causal language model pretrained exclusively on content relevant to the social sciences and humanities (SSH). It is intended as a domain-specific base model for SSH research and education, and is designed to be used through the MIRROR interface, which surfaces token-level surprisal rather than generating new text.
19
 
20
+ More information can be found in the paper [SHARE: Social-Humanities AI for Research and Education](https://huggingface.co/papers/2604.11152).
21
+
22
  **Note:** This is an intermediate checkpoint released after ~15% of planned pretraining (96B tokens of a target ~630B). It is a base (pretrained-only) model with no SFT, DPO, or RLHF. This base model is not suitable to chat applications.
23
 
24
  ## Model Details
 
36
  ### Model Sources
37
 
38
  - **Repository:** https://github.com/Joaoffg/SHARE
39
+ - **Paper:** [SHARE: Social-Humanities AI for Research and Education](https://arxiv.org/abs/2604.11152)
 
40
  - **Contact:** ferreiragoncalves@eshcc.eur.nl
41
 
42
  ## Uses