bakirgrbic commited on
Commit
6853176
·
verified ·
1 Parent(s): b36a071
Files changed (1) hide show
  1. README.md +46 -3
README.md CHANGED
@@ -1,3 +1,46 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ base_model:
6
+ - bsu-slim/electra-tiny
7
+ pipeline_tag: text-classification
8
+ library_name: transformers
9
+ ---
10
+
11
+ A pretrained [ELECTRA-Tiny](https://huggingface.co/bsu-slim/electra-tiny/tree/main) model. Pretraining [data](https://osf.io/5mk3x)
12
+ was from the [2024 BabyLM Challenge](https://babylm.github.io/index.html). Used personally to perform text classification
13
+ on the [Web of Science Dataset WOS-46985](https://data.mendeley.com/datasets/9rw3vkcfy4/6) but this model is not currently fine-tuned
14
+ for that task.
15
+
16
+
17
+ # Training
18
+ Used pretraining pipeline as defined in this [repository](https://github.com/bakirgrbic/bblm).
19
+
20
+ ## Hyperparameters
21
+ - Epochs: 1
22
+ - Batch size: 8
23
+ - Learning rate: 1e-4
24
+ - Optimizer: AdamW
25
+
26
+ ## Resources Used
27
+ - Compute: AWS Sagemaker ml.g4dn.xlarge
28
+ - Time: About 7 hours
29
+
30
+ # Evaluation (Web of Science)
31
+ Used wos pipeline as defined in this [repository](https://github.com/bakirgrbic/bblm).
32
+
33
+ ## Results
34
+ - 64% accuracy on the last epoch of the test set.
35
+
36
+ ## Hyperparameters
37
+ - Epochs: 3
38
+ - Batch size: 64
39
+ - Learning rate: 2e-5
40
+ - Optimizer: AdamW
41
+ - Max Length: 128
42
+ - Parameter Freezing: None
43
+
44
+ ## Resources Used
45
+ - Compute: AWS Sagemaker ml.g4dn.xlarge
46
+ - Time: About 5 minutes