DanielAvelar09 commited on
Commit
3599f46
·
verified ·
1 Parent(s): d08643d

End of training

Browse files
README.md CHANGED
@@ -1,11 +1,11 @@
1
  ---
 
 
2
  tags:
3
  - generated_from_trainer
4
  model-index:
5
  - name: mi_chatbot
6
  results: []
7
- language:
8
- - es
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -13,9 +13,9 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # mi_chatbot
15
 
16
- This model was trained from scratch on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 3.0442
19
 
20
  ## Model description
21
 
@@ -35,8 +35,8 @@ More information needed
35
 
36
  The following hyperparameters were used during training:
37
  - learning_rate: 3e-05
38
- - train_batch_size: 3
39
- - eval_batch_size: 3
40
  - seed: 42
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
@@ -46,18 +46,18 @@ The following hyperparameters were used during training:
46
 
47
  | Training Loss | Epoch | Step | Validation Loss |
48
  |:-------------:|:-----:|:----:|:---------------:|
49
- | No log | 1.0 | 37 | 3.8093 |
50
- | No log | 2.0 | 74 | 3.3021 |
51
- | No log | 3.0 | 111 | 3.0845 |
52
- | No log | 4.0 | 148 | 2.9462 |
53
- | No log | 5.0 | 185 | 2.8879 |
54
- | No log | 6.0 | 222 | 2.9819 |
55
- | No log | 7.0 | 259 | 3.0442 |
56
 
57
 
58
  ### Framework versions
59
 
60
- - Transformers 4.35.2
61
  - Pytorch 2.1.0+cu121
62
  - Datasets 2.17.1
63
- - Tokenizers 0.15.2
 
1
  ---
2
+ license: apache-2.0
3
+ base_model: mrm8488/distill-bert-base-spanish-wwm-cased-finetuned-spa-squad2-es
4
  tags:
5
  - generated_from_trainer
6
  model-index:
7
  - name: mi_chatbot
8
  results: []
 
 
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
13
 
14
  # mi_chatbot
15
 
16
+ This model is a fine-tuned version of [mrm8488/distill-bert-base-spanish-wwm-cased-finetuned-spa-squad2-es](https://huggingface.co/mrm8488/distill-bert-base-spanish-wwm-cased-finetuned-spa-squad2-es) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 0.5479
19
 
20
  ## Model description
21
 
 
35
 
36
  The following hyperparameters were used during training:
37
  - learning_rate: 3e-05
38
+ - train_batch_size: 8
39
+ - eval_batch_size: 8
40
  - seed: 42
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
 
46
 
47
  | Training Loss | Epoch | Step | Validation Loss |
48
  |:-------------:|:-----:|:----:|:---------------:|
49
+ | No log | 1.0 | 8 | 1.5199 |
50
+ | No log | 2.0 | 16 | 1.0632 |
51
+ | No log | 3.0 | 24 | 0.5938 |
52
+ | No log | 4.0 | 32 | 0.5494 |
53
+ | No log | 5.0 | 40 | 0.5706 |
54
+ | No log | 6.0 | 48 | 0.5299 |
55
+ | No log | 7.0 | 56 | 0.5479 |
56
 
57
 
58
  ### Framework versions
59
 
60
+ - Transformers 4.37.2
61
  - Pytorch 2.1.0+cu121
62
  - Datasets 2.17.1
63
+ - Tokenizers 0.15.2
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "mi_chatbot",
3
  "architectures": [
4
  "BertForQuestionAnswering"
5
  ],
@@ -19,7 +19,7 @@
19
  "pad_token_id": 1,
20
  "position_embedding_type": "absolute",
21
  "torch_dtype": "float32",
22
- "transformers_version": "4.35.2",
23
  "type_vocab_size": 2,
24
  "use_cache": true,
25
  "vocab_size": 31002
 
1
  {
2
+ "_name_or_path": "mrm8488/distill-bert-base-spanish-wwm-cased-finetuned-spa-squad2-es",
3
  "architectures": [
4
  "BertForQuestionAnswering"
5
  ],
 
19
  "pad_token_id": 1,
20
  "position_embedding_type": "absolute",
21
  "torch_dtype": "float32",
22
+ "transformers_version": "4.37.2",
23
  "type_vocab_size": 2,
24
  "use_cache": true,
25
  "vocab_size": 31002
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b4dc30c35f08dcdc6e9d152bfc16aa49674cefecbd73afd7faa7bc6cdf4bcaad
3
  size 437070648
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10f0b640da5c1ad66bd065efc82208cf25ed3324ebd1d09b6c6113c7bd560f05
3
  size 437070648
runs/Feb24_20-36-28_ed064940385f/events.out.tfevents.1708806990.ed064940385f.1370.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9342400a35d4b887a30d13feb992328dd873ebcea40f509a5e28660726b15858
3
+ size 6581
special_tokens_map.json CHANGED
@@ -1,37 +1,7 @@
1
  {
2
- "cls_token": {
3
- "content": "[CLS]",
4
- "lstrip": false,
5
- "normalized": false,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "mask_token": {
10
- "content": "[MASK]",
11
- "lstrip": false,
12
- "normalized": false,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "[PAD]",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
- "sep_token": {
24
- "content": "[SEP]",
25
- "lstrip": false,
26
- "normalized": false,
27
- "rstrip": false,
28
- "single_word": false
29
- },
30
- "unk_token": {
31
- "content": "[UNK]",
32
- "lstrip": false,
33
- "normalized": false,
34
- "rstrip": false,
35
- "single_word": false
36
- }
37
  }
 
1
  {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c4056a4d5a06c624704768773c8dc416691a520cc01120993ab138124d4ea9b
3
- size 729903
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b2cb0a433da31658bd79f95e7beabf22a59aaaf0d39513b24f5474cef8edcfd
3
+ size 730166
tokenizer_config.json CHANGED
@@ -46,19 +46,12 @@
46
  "do_basic_tokenize": true,
47
  "do_lower_case": true,
48
  "mask_token": "[MASK]",
49
- "max_length": 384,
50
  "model_max_length": 1000000000000000019884624838656,
51
  "never_split": null,
52
- "pad_to_multiple_of": null,
53
  "pad_token": "[PAD]",
54
- "pad_token_type_id": 0,
55
- "padding_side": "right",
56
  "sep_token": "[SEP]",
57
- "stride": 0,
58
  "strip_accents": null,
59
  "tokenize_chinese_chars": true,
60
  "tokenizer_class": "BertTokenizer",
61
- "truncation_side": "right",
62
- "truncation_strategy": "only_second",
63
  "unk_token": "[UNK]"
64
  }
 
46
  "do_basic_tokenize": true,
47
  "do_lower_case": true,
48
  "mask_token": "[MASK]",
 
49
  "model_max_length": 1000000000000000019884624838656,
50
  "never_split": null,
 
51
  "pad_token": "[PAD]",
 
 
52
  "sep_token": "[SEP]",
 
53
  "strip_accents": null,
54
  "tokenize_chinese_chars": true,
55
  "tokenizer_class": "BertTokenizer",
 
 
56
  "unk_token": "[UNK]"
57
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:76aba73107fc0595dcf4b3c219f0c566f2504e5aeb6b738b662ae66f6f1eab51
3
- size 4536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a55afa51176ba0c93b2b9674af553a0ec406f291f5c3c50183cdc1739f6fc7a
3
+ size 4664