deuswoof commited on
Commit
00a31ec
·
1 Parent(s): 399d2a7

Training in progress, step 10

Browse files
25_10_23_config_test_2.csv ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ run_number,comment,peformed_already,num_train_epochs,max_tokens,temperature,stop_token,classification_of_valuems,stemming,lemmatization
2
+ 1,no variations,False,2,100,0.8,False,False,False,False
3
+ 2,lemmatization set True,False,2,100,0.8,False,False,False,True
4
+ 3,stemming set True,False,2,100,0.8,False,False,True,False
5
+ 4,classification_of_valuems set True,False,2,100,0.8,False,True,False,False
6
+ 5,"lemmatization set True
7
+ stemming set True
8
+ classification_of_valuems set True",False,2,100,0.8,False,True,True,True
9
+ 6,"lemmatization set True
10
+ classification_of_valuems set True",False,2,100,0.8,False,True,False,True
11
+ 7,"lemmatization set True
12
+ stemming set True
13
+ classification_of_valuems set True",False,2,100,0.8,False,True,True,True
25_10_23_results_real.csv CHANGED
@@ -11,4 +11,4 @@ run_number,items_per_minute,changed_settings,total_time_taken,rouge_scores_unnes
11
  10,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
12
  11,1521.609931137608,classification_of_valuems set True,17.35004448890686,0,0.3097912503061737,0.3117918619129387,0.2985123093297854,0.3692864464982227,0.3657660910086977,0.3510613139478704,0.4322932016041544,0.4194092516901677,0.4034142346670022,0.1853986019402615,0.1767452335316192,0.1744893735961138,0.2321160496220204,0.2189556994085239,0.2145193605708969,0.2770436971412456,0.2594983484442184,0.2537992371891699
13
  12,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
14
- 13,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
 
11
  10,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
12
  11,1521.609931137608,classification_of_valuems set True,17.35004448890686,0,0.3097912503061737,0.3117918619129387,0.2985123093297854,0.3692864464982227,0.3657660910086977,0.3510613139478704,0.4322932016041544,0.4194092516901677,0.4034142346670022,0.1853986019402615,0.1767452335316192,0.1744893735961138,0.2321160496220204,0.2189556994085239,0.2145193605708969,0.2770436971412456,0.2594983484442184,0.2537992371891699
13
  12,0.0,0,0.0,0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0,0.0
14
+ 13,1521.0856607375722,0,17.356024503707886,0,0.2370841300636563,0.304945047380962,0.2519327504848678,0.2724776461383102,0.3442978964808718,0.2880887444999471,0.3105848069814388,0.382543974390498,0.3224917625996572,0.0799995605154184,0.0919100818000075,0.0819151637902611,0.0992904624866073,0.1133134464313628,0.1005193519003102,0.1205682396998857,0.1386388830965476,0.12211284514134889
25_10_23_results_real_1.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ run_number,items_per_minute,changed_settings,total_time_taken,rouge_scores_unnest,rouge1 low Precision,rouge1 low Recall,rouge1 low F1 Score,rouge1 mid Precision,rouge1 mid Recall,rouge1 mid F1 Score,rouge1 high Precision,rouge1 high Recall,rouge1 high F1 Score,rouge2 low Precision,rouge2 low Recall,rouge2 low F1 Score,rouge2 mid Precision,rouge2 mid Recall,rouge2 mid F1 Score,rouge2 high Precision,rouge2 high Recall,rouge2 high F1 Score,min_cosine_sim_value,max_cosine_sim_value,mean_cosine_sim_value
2
+ 1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
3
+ 2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
README.md CHANGED
@@ -628,6 +628,18 @@ The following `bitsandbytes` quantization config was used during training:
628
  - bnb_4bit_use_double_quant: True
629
  - bnb_4bit_compute_dtype: bfloat16
630
 
 
 
 
 
 
 
 
 
 
 
 
 
631
  The following `bitsandbytes` quantization config was used during training:
632
  - quant_method: bitsandbytes
633
  - load_in_8bit: False
@@ -693,5 +705,6 @@ The following `bitsandbytes` quantization config was used during training:
693
  - PEFT 0.5.0
694
  - PEFT 0.5.0
695
  - PEFT 0.5.0
 
696
 
697
  - PEFT 0.5.0
 
628
  - bnb_4bit_use_double_quant: True
629
  - bnb_4bit_compute_dtype: bfloat16
630
 
631
+ The following `bitsandbytes` quantization config was used during training:
632
+ - quant_method: bitsandbytes
633
+ - load_in_8bit: False
634
+ - load_in_4bit: True
635
+ - llm_int8_threshold: 6.0
636
+ - llm_int8_skip_modules: None
637
+ - llm_int8_enable_fp32_cpu_offload: False
638
+ - llm_int8_has_fp16_weight: False
639
+ - bnb_4bit_quant_type: nf4
640
+ - bnb_4bit_use_double_quant: True
641
+ - bnb_4bit_compute_dtype: bfloat16
642
+
643
  The following `bitsandbytes` quantization config was used during training:
644
  - quant_method: bitsandbytes
645
  - load_in_8bit: False
 
705
  - PEFT 0.5.0
706
  - PEFT 0.5.0
707
  - PEFT 0.5.0
708
+ - PEFT 0.5.0
709
 
710
  - PEFT 0.5.0
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f70e4f16e44af1d9575ca763b176ef77a05f522543ad516aa514b5f821635d20
3
  size 100733709
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39cee32fdfdf77a10bbb2f76b2e455c5b9f50213de36c364a7673f63c031cbd2
3
  size 100733709
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5a3480b947478c17d5df359b02a08abb839a59e985f611c4c15b2f0b3a0cb324
3
  size 100690288
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f24a31bf80360e505636d8424b9931cff01274b92ecaedbbd0590321fd681d6a
3
  size 100690288
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:91c44e8dc8169fffdaa1ca9accb53d71deab0240c8b8ce38c8f9a2dcce55449d
3
  size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dde72a63056d670c29f80ad8c5b9aa6efa0445a1d2e70853d26f5091133822bd
3
  size 4283