zsomai commited on
Commit
032c862
1 Parent(s): bfa9735

Training in progress epoch 0

Browse files
Files changed (4) hide show
  1. README.md +2 -2
  2. tf_model.h5 +1 -1
  3. tokenizer.json +2 -16
  4. tokenizer_config.json +1 -1
README.md CHANGED
@@ -14,7 +14,7 @@ probably proofread and complete it, then remove this comment. -->
14
 
15
  This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
- - Train Loss: 0.6832
18
  - Epoch: 0
19
 
20
  ## Model description
@@ -41,7 +41,7 @@ The following hyperparameters were used during training:
41
 
42
  | Train Loss | Epoch |
43
  |:----------:|:-----:|
44
- | 0.6832 | 0 |
45
 
46
 
47
  ### Framework versions
 
14
 
15
  This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Train Loss: 0.7777
18
  - Epoch: 0
19
 
20
  ## Model description
 
41
 
42
  | Train Loss | Epoch |
43
  |:----------:|:-----:|
44
+ | 0.7777 | 0 |
45
 
46
 
47
  ### Framework versions
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d6bee7c9c90abab78cb181d5c2b72c4d14414714caf3ed026d2c52a1671ca5a2
3
  size 431154128
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd66e742c2fd924b27c2be4ece19e920e99a9bb04ca3f8e44153384130adc6bd
3
  size 431154128
tokenizer.json CHANGED
@@ -1,21 +1,7 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 384,
6
- "strategy": "OnlySecond",
7
- "stride": 128
8
- },
9
- "padding": {
10
- "strategy": {
11
- "Fixed": 384
12
- },
13
- "direction": "Right",
14
- "pad_to_multiple_of": null,
15
- "pad_id": 0,
16
- "pad_type_id": 0,
17
- "pad_token": "[PAD]"
18
- },
19
  "added_tokens": [
20
  {
21
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
tokenizer_config.json CHANGED
@@ -8,6 +8,6 @@
8
  "sep_token": "[SEP]",
9
  "strip_accents": null,
10
  "tokenize_chinese_chars": true,
11
- "tokenizer_class": "BertTokenizer",
12
  "unk_token": "[UNK]"
13
  }
 
8
  "sep_token": "[SEP]",
9
  "strip_accents": null,
10
  "tokenize_chinese_chars": true,
11
+ "tokenizer_class": "DistilBertTokenizer",
12
  "unk_token": "[UNK]"
13
  }