EternalRecursion commited on
Commit
6a8b5b9
·
1 Parent(s): 90ce9d8

Training in progress epoch 0

Browse files
README.md CHANGED
@@ -12,11 +12,11 @@ probably proofread and complete it, then remove this comment. -->
12
 
13
  # EternalRecursion/roberta-finetuned-poetry
14
 
15
- This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
- - Train Loss: 1.9830
18
- - Validation Loss: 1.8248
19
- - Epoch: 2
20
 
21
  ## Model description
22
 
@@ -42,9 +42,7 @@ The following hyperparameters were used during training:
42
 
43
  | Train Loss | Validation Loss | Epoch |
44
  |:----------:|:---------------:|:-----:|
45
- | 2.1615 | 1.8898 | 0 |
46
- | 2.0215 | 1.8530 | 1 |
47
- | 1.9830 | 1.8248 | 2 |
48
 
49
 
50
  ### Framework versions
 
12
 
13
  # EternalRecursion/roberta-finetuned-poetry
14
 
15
+ This model is a fine-tuned version of [roberta-large](https://huggingface.co/roberta-large) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Train Loss: 2.0071
18
+ - Validation Loss: 1.7287
19
+ - Epoch: 0
20
 
21
  ## Model description
22
 
 
42
 
43
  | Train Loss | Validation Loss | Epoch |
44
  |:----------:|:---------------:|:-----:|
45
+ | 2.0071 | 1.7287 | 0 |
 
 
46
 
47
 
48
  ### Framework versions
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "roberta-base",
3
  "architectures": [
4
  "RobertaForMaskedLM"
5
  ],
@@ -9,14 +9,14 @@
9
  "eos_token_id": 2,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
- "hidden_size": 768,
13
  "initializer_range": 0.02,
14
- "intermediate_size": 3072,
15
  "layer_norm_eps": 1e-05,
16
  "max_position_embeddings": 514,
17
  "model_type": "roberta",
18
- "num_attention_heads": 12,
19
- "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
  "transformers_version": "4.28.1",
 
1
  {
2
+ "_name_or_path": "roberta-large",
3
  "architectures": [
4
  "RobertaForMaskedLM"
5
  ],
 
9
  "eos_token_id": 2,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 1024,
13
  "initializer_range": 0.02,
14
+ "intermediate_size": 4096,
15
  "layer_norm_eps": 1e-05,
16
  "max_position_embeddings": 514,
17
  "model_type": "roberta",
18
+ "num_attention_heads": 16,
19
+ "num_hidden_layers": 24,
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
  "transformers_version": "4.28.1",
logs/train/events.out.tfevents.1682888132.t1v-n-54f58e78-w-0.75868.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fefbe0549c04077ca1c9fe5d82847584ecbf0850fd6c6dfec940df9b166b14a6
3
+ size 6899234
logs/validation/events.out.tfevents.1682888749.t1v-n-54f58e78-w-0.75868.1.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92f8a8612bffee7334b81f8d83f93332e7179158a15d43f7d84d834dbc4972fe
3
+ size 232
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca33c1f167ff837cf9943d9e7451ca95f63c6d5044281718899715f0ab83c0f6
3
- size 655071156
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6015426abbe3ddccc7096a1ed433aa0ce3145280488134f34ef72d88d28476e9
3
+ size 1630210764