bakhareva commited on
Commit
640b214
·
verified ·
1 Parent(s): cc5e240

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "bert-base-uncased",
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
 
1
  {
2
+ "_name_or_path": "ProsusAI/finbert",
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e091cc1987727331a62405da8d51f54cc6f0e66f7a97a064c94da60ece69c9b6
3
  size 437958648
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1074aa7892e16b4fc2c07bfeba5a3a4c43a93abfb3d148fbd947752579db19d
3
  size 437958648
runs/Jul12_12-28-08_cd2cda595b0e/events.out.tfevents.1720787298.cd2cda595b0e.516.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53ed3ec2da2f79cffaf28ea5af6393819c667606a1e1fcc2095e6c4e90ebaf5d
3
+ size 5327
tokenizer_config.json CHANGED
@@ -43,9 +43,11 @@
43
  },
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
 
46
  "do_lower_case": true,
47
  "mask_token": "[MASK]",
48
  "model_max_length": 512,
 
49
  "pad_token": "[PAD]",
50
  "sep_token": "[SEP]",
51
  "strip_accents": null,
 
43
  },
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
  "do_lower_case": true,
48
  "mask_token": "[MASK]",
49
  "model_max_length": 512,
50
+ "never_split": null,
51
  "pad_token": "[PAD]",
52
  "sep_token": "[SEP]",
53
  "strip_accents": null,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b721a86794eda9d83d33fc68bd98645517d57a8b0e45af9f42ed3c9fd79e40a3
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90928b7d27518bce6b281f08230016f48fac67d93628aa90a0855b4ea68eae98
3
  size 5176