mongdiutindei commited on
Commit
c0e7f92
1 Parent(s): 8676748

Training complete

Browse files
README.md CHANGED
@@ -3,6 +3,8 @@ license: apache-2.0
3
  base_model: bert-base-cased
4
  tags:
5
  - generated_from_trainer
 
 
6
  model-index:
7
  - name: bert-finetuned-squad
8
  results: []
@@ -13,7 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # bert-finetuned-squad
15
 
16
- This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the None dataset.
17
 
18
  ## Model description
19
 
@@ -32,13 +34,13 @@ More information needed
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
35
- - learning_rate: 0.001
36
  - train_batch_size: 8
37
  - eval_batch_size: 8
38
  - seed: 42
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: linear
41
- - num_epochs: 1
42
 
43
  ### Training results
44
 
@@ -46,7 +48,7 @@ The following hyperparameters were used during training:
46
 
47
  ### Framework versions
48
 
49
- - Transformers 4.34.1
50
  - Pytorch 2.1.0+cu118
51
  - Datasets 2.14.6
52
  - Tokenizers 0.14.1
 
3
  base_model: bert-base-cased
4
  tags:
5
  - generated_from_trainer
6
+ datasets:
7
+ - squad
8
  model-index:
9
  - name: bert-finetuned-squad
10
  results: []
 
15
 
16
  # bert-finetuned-squad
17
 
18
+ This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the squad dataset.
19
 
20
  ## Model description
21
 
 
34
  ### Training hyperparameters
35
 
36
  The following hyperparameters were used during training:
37
+ - learning_rate: 0.0001
38
  - train_batch_size: 8
39
  - eval_batch_size: 8
40
  - seed: 42
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
+ - num_epochs: 20
44
 
45
  ### Training results
46
 
 
48
 
49
  ### Framework versions
50
 
51
+ - Transformers 4.35.0
52
  - Pytorch 2.1.0+cu118
53
  - Datasets 2.14.6
54
  - Tokenizers 0.14.1
config.json CHANGED
@@ -19,7 +19,7 @@
19
  "pad_token_id": 0,
20
  "position_embedding_type": "absolute",
21
  "torch_dtype": "float32",
22
- "transformers_version": "4.34.1",
23
  "type_vocab_size": 2,
24
  "use_cache": true,
25
  "vocab_size": 28996
 
19
  "pad_token_id": 0,
20
  "position_embedding_type": "absolute",
21
  "torch_dtype": "float32",
22
+ "transformers_version": "4.35.0",
23
  "type_vocab_size": 2,
24
  "use_cache": true,
25
  "vocab_size": 28996
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6af22663f976e3c51d4a95c0d0d7d8a6c9e7ebae08856c94c08ac7c271524bf
3
+ size 430908208
runs/Nov03_07-38-49_1e3ac9194a54/events.out.tfevents.1698997138.1e3ac9194a54.170.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2bf01816fc16c0c97624c4d81b3102f85931cb84de7c5fc8a1a31a7aa0dca3c
3
+ size 4564
runs/Nov03_07-43-37_1e3ac9194a54/events.out.tfevents.1698997418.1e3ac9194a54.170.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07efd3607374c0f5d9e639954f2e58554a2bad4600a046729223dc6a9e7ae5a5
3
+ size 5388
tokenizer.json CHANGED
@@ -1,21 +1,7 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 384,
6
- "strategy": "OnlySecond",
7
- "stride": 128
8
- },
9
- "padding": {
10
- "strategy": {
11
- "Fixed": 384
12
- },
13
- "direction": "Right",
14
- "pad_to_multiple_of": null,
15
- "pad_id": 0,
16
- "pad_type_id": 0,
17
- "pad_token": "[PAD]"
18
- },
19
  "added_tokens": [
20
  {
21
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a331d5c205afe6ad50655cc3bbcf6651d9b3a0a429f0c64f07f4e56ba4077f79
3
- size 4536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:245f9b9f7e397bffa39b213a939d7a678d4c128bd33431548ca616b3827e5b35
3
+ size 4600