ayshi commited on
Commit
f2b8e3e
·
1 Parent(s): f3a3f09

Training in progress epoch 0

Browse files
Files changed (5) hide show
  1. README.md +8 -17
  2. config.json +1 -1
  3. special_tokens_map.json +0 -7
  4. tf_model.h5 +2 -2
  5. tokenizer_config.json +1 -11
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  license: apache-2.0
3
- base_model: ayshi/basic_distil
4
  tags:
5
  - generated_from_keras_callback
6
  model-index:
@@ -13,12 +13,12 @@ probably proofread and complete it, then remove this comment. -->
13
 
14
  # ayshi/basic_distil
15
 
16
- This model is a fine-tuned version of [ayshi/basic_distil](https://huggingface.co/ayshi/basic_distil) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Train Loss: 0.0069
19
- - Validation Loss: 1.3626
20
- - Train Accuracy: 0.7556
21
- - Epoch: 9
22
 
23
  ## Model description
24
 
@@ -44,16 +44,7 @@ The following hyperparameters were used during training:
44
 
45
  | Train Loss | Validation Loss | Train Accuracy | Epoch |
46
  |:----------:|:---------------:|:--------------:|:-----:|
47
- | 0.0198 | 1.1937 | 0.7911 | 0 |
48
- | 0.0220 | 1.3049 | 0.7511 | 1 |
49
- | 0.0186 | 1.5090 | 0.7822 | 2 |
50
- | 0.0181 | 1.3297 | 0.7867 | 3 |
51
- | 0.0095 | 1.4989 | 0.68 | 4 |
52
- | 0.0101 | 1.4034 | 0.7689 | 5 |
53
- | 0.0076 | 1.3720 | 0.7689 | 6 |
54
- | 0.0077 | 1.3592 | 0.76 | 7 |
55
- | 0.0065 | 1.3647 | 0.7556 | 8 |
56
- | 0.0069 | 1.3626 | 0.7556 | 9 |
57
 
58
 
59
  ### Framework versions
@@ -61,4 +52,4 @@ The following hyperparameters were used during training:
61
  - Transformers 4.34.0
62
  - TensorFlow 2.13.0
63
  - Datasets 2.14.5
64
- - Tokenizers 0.14.0
 
1
  ---
2
  license: apache-2.0
3
+ base_model: distilbert-base-multilingual-cased
4
  tags:
5
  - generated_from_keras_callback
6
  model-index:
 
13
 
14
  # ayshi/basic_distil
15
 
16
+ This model is a fine-tuned version of [distilbert-base-multilingual-cased](https://huggingface.co/distilbert-base-multilingual-cased) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Train Loss: 1.4022
19
+ - Validation Loss: 1.1513
20
+ - Train Accuracy: 0.6667
21
+ - Epoch: 0
22
 
23
  ## Model description
24
 
 
44
 
45
  | Train Loss | Validation Loss | Train Accuracy | Epoch |
46
  |:----------:|:---------------:|:--------------:|:-----:|
47
+ | 1.4022 | 1.1513 | 0.6667 | 0 |
 
 
 
 
 
 
 
 
 
48
 
49
 
50
  ### Framework versions
 
52
  - Transformers 4.34.0
53
  - TensorFlow 2.13.0
54
  - Datasets 2.14.5
55
+ - Tokenizers 0.14.1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "ayshi/basic_distil",
3
  "activation": "gelu",
4
  "architectures": [
5
  "DistilBertForSequenceClassification"
 
1
  {
2
+ "_name_or_path": "distilbert-base-multilingual-cased",
3
  "activation": "gelu",
4
  "architectures": [
5
  "DistilBertForSequenceClassification"
special_tokens_map.json CHANGED
@@ -1,11 +1,4 @@
1
  {
2
- "additional_special_tokens": [
3
- "[PAD]",
4
- "[UNK]",
5
- "[CLS]",
6
- "[SEP]",
7
- "[MASK]"
8
- ],
9
  "cls_token": "[CLS]",
10
  "mask_token": "[MASK]",
11
  "pad_token": "[PAD]",
 
1
  {
 
 
 
 
 
 
 
2
  "cls_token": "[CLS]",
3
  "mask_token": "[MASK]",
4
  "pad_token": "[PAD]",
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a21e7a86ad0ab691d73fa3af4989cb2f43c79cfeb5dc30851ca4888634a8587e
3
- size 541452232
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8dc816525040da74b29612a32b721a7383fcc1ba2b328b441ed58bd7f0e2f282
3
+ size 541448896
tokenizer_config.json CHANGED
@@ -41,26 +41,16 @@
41
  "special": true
42
  }
43
  },
44
- "additional_special_tokens": [
45
- "[PAD]",
46
- "[UNK]",
47
- "[CLS]",
48
- "[SEP]",
49
- "[MASK]"
50
- ],
51
  "clean_up_tokenization_spaces": true,
52
  "cls_token": "[CLS]",
53
  "do_lower_case": false,
54
  "mask_token": "[MASK]",
55
- "max_length": 512,
56
  "model_max_length": 512,
57
  "pad_token": "[PAD]",
58
  "sep_token": "[SEP]",
59
- "stride": 0,
60
  "strip_accents": null,
61
  "tokenize_chinese_chars": true,
62
  "tokenizer_class": "DistilBertTokenizer",
63
- "truncation_side": "right",
64
- "truncation_strategy": "longest_first",
65
  "unk_token": "[UNK]"
66
  }
 
41
  "special": true
42
  }
43
  },
44
+ "additional_special_tokens": [],
 
 
 
 
 
 
45
  "clean_up_tokenization_spaces": true,
46
  "cls_token": "[CLS]",
47
  "do_lower_case": false,
48
  "mask_token": "[MASK]",
 
49
  "model_max_length": 512,
50
  "pad_token": "[PAD]",
51
  "sep_token": "[SEP]",
 
52
  "strip_accents": null,
53
  "tokenize_chinese_chars": true,
54
  "tokenizer_class": "DistilBertTokenizer",
 
 
55
  "unk_token": "[UNK]"
56
  }