duraad commited on
Commit
56a2d6c
·
verified ·
1 Parent(s): dfcb9b3

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "dura-garage/nepberta2nepberta",
3
  "architectures": [
4
  "EncoderDecoderModel"
5
  ],
@@ -166,6 +166,7 @@
166
  "use_cache": true,
167
  "vocab_size": 30522
168
  },
 
169
  "from_tf": true,
170
  "is_encoder_decoder": true,
171
  "model_type": "encoder-decoder",
 
1
  {
2
+ "_name_or_path": "duraad/finetuning-b2b",
3
  "architectures": [
4
  "EncoderDecoderModel"
5
  ],
 
166
  "use_cache": true,
167
  "vocab_size": 30522
168
  },
169
+ "eos_token_id": 2,
170
  "from_tf": true,
171
  "is_encoder_decoder": true,
172
  "model_type": "encoder-decoder",
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1fc579c1f4f74e953ef7588b7375fdc36d530986b929641db7c4e09e36d3ff02
3
  size 989518784
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8226832367d5d31e78d22e736a576384f404b65781dc083cfbce302164445ecb
3
  size 989518784
special_tokens_map.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "cls_token": {
3
  "content": "[CLS]",
4
  "lstrip": false,
@@ -6,6 +7,7 @@
6
  "rstrip": false,
7
  "single_word": false
8
  },
 
9
  "mask_token": {
10
  "content": "[MASK]",
11
  "lstrip": false,
 
1
  {
2
+ "bos_token": "[CLS]",
3
  "cls_token": {
4
  "content": "[CLS]",
5
  "lstrip": false,
 
7
  "rstrip": false,
8
  "single_word": false
9
  },
10
+ "eos_token": "[SEP]",
11
  "mask_token": {
12
  "content": "[MASK]",
13
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -41,17 +41,26 @@
41
  "special": true
42
  }
43
  },
 
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
  "do_basic_tokenize": true,
47
  "do_lower_case": true,
 
48
  "mask_token": "[MASK]",
 
49
  "model_max_length": 1000000000000000019884624838656,
50
  "never_split": null,
 
51
  "pad_token": "[PAD]",
 
 
52
  "sep_token": "[SEP]",
 
53
  "strip_accents": null,
54
  "tokenize_chinese_chars": true,
55
  "tokenizer_class": "BertTokenizer",
 
 
56
  "unk_token": "[UNK]"
57
  }
 
41
  "special": true
42
  }
43
  },
44
+ "bos_token": "[CLS]",
45
  "clean_up_tokenization_spaces": true,
46
  "cls_token": "[CLS]",
47
  "do_basic_tokenize": true,
48
  "do_lower_case": true,
49
+ "eos_token": "[SEP]",
50
  "mask_token": "[MASK]",
51
+ "max_length": 512,
52
  "model_max_length": 1000000000000000019884624838656,
53
  "never_split": null,
54
+ "pad_to_multiple_of": null,
55
  "pad_token": "[PAD]",
56
+ "pad_token_type_id": 0,
57
+ "padding_side": "right",
58
  "sep_token": "[SEP]",
59
+ "stride": 0,
60
  "strip_accents": null,
61
  "tokenize_chinese_chars": true,
62
  "tokenizer_class": "BertTokenizer",
63
+ "truncation_side": "right",
64
+ "truncation_strategy": "longest_first",
65
  "unk_token": "[UNK]"
66
  }