cglez commited on
Commit
3dd4f6b
·
verified ·
1 Parent(s): 0df7b25

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +60 -39
README.md CHANGED
@@ -2,72 +2,93 @@
2
  library_name: transformers
3
  language: en
4
  license: apache-2.0
5
- datasets: []
6
- tags: []
 
 
7
  ---
8
 
9
- # Model Card for <Model>
10
 
11
- A pretrained BERT using <Dataset>.
12
 
13
  ## Model Details
14
 
15
- ### Model Description
16
 
17
- A MLM-only pretrained BERT-base using <Dataset>.
 
 
18
 
19
  - **Developed by:** [Cesar Gonzalez-Gutierrez](https://ceguel.es)
20
  - **Funded by:** [ERC](https://erc.europa.eu)
21
- - **Model type:** MLM pretrained BERT
22
- - **Language(s) (NLP):** English
23
- - **License:** Apache license 2.0
24
- - **Pretrained from model:** [BERT base model (uncased)](https://huggingface.co/google-bert/bert-base-uncased)
25
-
26
- ### Model Checkpoints
27
-
28
- [More Information Needed]
29
-
30
- ### Model Sources
31
-
32
- - **Paper:** [More Information Needed]
33
-
34
- ## Uses
35
-
36
- See <https://huggingface.co/google-bert/bert-base-uncased#intended-uses--limitations>.
37
-
38
- ### Checkpoint Use
39
-
40
- [More Information Needed]
41
-
42
- ## Bias, Risks, and Limitations
43
-
44
- See <https://huggingface.co/google-bert/bert-base-uncased#limitations-and-bias>.
 
 
 
 
 
 
 
 
 
 
 
 
45
 
46
  ## Training Details
47
 
48
- See <https://huggingface.co/google-bert/bert-base-uncased#training-procedure>.
 
49
 
50
  ### Training Data
51
 
52
- [More Information Needed]
53
-
54
- #### Preprocessing [optional]
55
-
56
- [More Information Needed]
57
 
58
  #### Training Hyperparameters
59
 
60
- - **Training regime:** fp16
61
  - **Batch size:** 32
62
  - **Gradient accumulation steps:** 3
63
 
 
 
 
 
 
 
 
 
 
 
64
  ## Environmental Impact
65
 
66
  - **Hardware Type:** NVIDIA Tesla V100 PCIE 32GB
67
- - **Hours used:** [More Information Needed]
68
  - **Cluster Provider:** [Artemisa](https://artemisa.ific.uv.es/web/)
69
  - **Compute Region:** EU
70
- - **Carbon Emitted:** [More Information Needed] <!-- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). -->
71
 
72
  ## Citation
73
 
 
2
  library_name: transformers
3
  language: en
4
  license: apache-2.0
5
+ datasets:
6
+ - CogComp/trec
7
+ base_model:
8
+ - google-bert/bert-base-uncased
9
  ---
10
 
11
+ # Model Card: BERT-DAPT-TREC
12
 
13
+ A domain-adapted BERT-base model, further pre-trained on the TREC dataset text.
14
 
15
  ## Model Details
16
 
17
+ ### Description
18
 
19
+ This model is based on the [BERT base (uncased)](https://huggingface.co/google-bert/bert-base-uncased)
20
+ architecture and was further pre-trained (domain-adapted) using the text in TREC dataset, excluding its test split.
21
+ Only the masked language modeling (MLM) objective was used during domain adaptation.
22
 
23
  - **Developed by:** [Cesar Gonzalez-Gutierrez](https://ceguel.es)
24
  - **Funded by:** [ERC](https://erc.europa.eu)
25
+ - **Architecture:** BERT-base
26
+ - **Language:** English
27
+ - **License:** Apache 2.0
28
+ - **Base model:** [BERT base model (uncased)](https://huggingface.co/google-bert/bert-base-uncased)
29
+
30
+ ### Checkpoints
31
+
32
+ Intermediate checkpoints from the pre-training process are available and can be accessed using specific tags,
33
+ which correspond to training epochs and steps:
34
+
35
+ | Epoch | Step | Tags | |
36
+ |---|---|---|---|
37
+ | 1 | 51 | epoch-1 | step-51 |
38
+ | 5 | 256 | epoch-5 | step-256 |
39
+ | 10 | 513 | epoch-10 | step-513 |
40
+ | 20 | 1026 | epoch-20 | step-1026 |
41
+ | 40 | 2053 | epoch-40 | step-2053 |
42
+ | 60 | 3080 | epoch-60 | step-3080 |
43
+ | 80 | 4106 | epoch-80 | step-4106 |
44
+ | 99 | 5100 | epoch-99 | step-5100 |
45
+ | 120 | 6126 | epoch-120 | step-6126 |
46
+ | 140 | 7153 | epoch-140 | step-7153 |
47
+ | 160 | 8180 | epoch-160 | step-8180 |
48
+ | 180 | 9206 | epoch-180 | step-9206 |
49
+ | 199 | 10200 | epoch-199 | step-10200 |
50
+
51
+ To load a model from a specific intermediate checkpoint, use the `revision` parameter with the corresponding tag:
52
+ ```python
53
+ from transformers import AutoModelForMaskedLM
54
+
55
+ model = AutoModelForMaskedLM.from_pretrained("<model-name>", revision="<checkpoint-tag>")
56
+ ```
57
+
58
+ ### Sources
59
+
60
+ - **Paper:** [Information pending]
61
 
62
  ## Training Details
63
 
64
+ For more details on the training procedure, please refer to the base model's documentation:
65
+ [Training procedure](https://huggingface.co/google-bert/bert-base-uncased#training-procedure).
66
 
67
  ### Training Data
68
 
69
+ All texts from TREC dataset, excluding the test partition.
 
 
 
 
70
 
71
  #### Training Hyperparameters
72
 
73
+ - **Precision:** fp16
74
  - **Batch size:** 32
75
  - **Gradient accumulation steps:** 3
76
 
77
+ ## Uses
78
+
79
+ For typical use cases and limitations, please refer to the base model's guidance:
80
+ [Inteded uses & limitations](https://huggingface.co/google-bert/bert-base-uncased#intended-uses--limitations).
81
+
82
+ ## Bias, Risks, and Limitations
83
+
84
+ This model inherits potential risks and limitations from the base model. Refer to:
85
+ [Limitations and bias](https://huggingface.co/google-bert/bert-base-uncased#limitations-and-bias).
86
+
87
  ## Environmental Impact
88
 
89
  - **Hardware Type:** NVIDIA Tesla V100 PCIE 32GB
 
90
  - **Cluster Provider:** [Artemisa](https://artemisa.ific.uv.es/web/)
91
  - **Compute Region:** EU
 
92
 
93
  ## Citation
94