Ishaank18 commited on
Commit
7c816c2
·
verified ·
1 Parent(s): 77f95a0

Upload folder using huggingface_hub

Browse files
MoE-Hash_human_eval_samples.txt ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ================================================================================
3
+ SAMPLE 1 (ROUGE-1: 0.000)
4
+ ================================================================================
5
+
6
+ DOCUMENT (first 300 characters):
7
+ Two snowsports enthusiasts got married at a Scottish ski resort before sliding off down a run in their wedding attire.
8
+ Bridget and Jonathan Reid, from Moy, near Tomatin in the Highlands, tied the knot at Nevis Range, near Fort William, on Friday.
9
+ The couple first's date six years ago was a skiing tr...
10
+
11
+ REFERENCE SUMMARY:
12
+ All images copyrighted.
13
+
14
+ GENERATED SUMMARY:
15
+ A couple of years ago have been married on a wedding dress at Nevis Range.
16
+
17
+ EVALUATION (fill in your analysis):
18
+ 1. Content Relevance (Does it capture main points?):
19
+ [Your evaluation: Score 1-5, explanation]
20
+
21
+ 2. Coherence (Does it flow well?):
22
+ [Your evaluation: Score 1-5, explanation]
23
+
24
+ 3. Fluency (Is it grammatically correct?):
25
+ [Your evaluation: Score 1-5, explanation]
26
+
27
+ 4. Factual Consistency (Are facts accurate?):
28
+ [Your evaluation: Score 1-5, explanation]
29
+
30
+ OVERALL SCORE: [X/20]
31
+
32
+ ================================================================================
33
+
34
+ ================================================================================
35
+ SAMPLE 2 (ROUGE-1: 0.286)
36
+ ================================================================================
37
+
38
+ DOCUMENT (first 300 characters):
39
+ Mandla Hlatshwayo and his friend were shot after confronting a group of men who had robbed women of their mobile phones in a pub in Soweto.
40
+ The suspects were found in possession of drugs and an unlicensed gun.
41
+ South Africa has one of the highest murder rates in the world with more than 50,000 cases ...
42
+
43
+ REFERENCE SUMMARY:
44
+ South African police say four people have been arrested in connection with the murder of former actor on popular local TV series Generations.
45
+
46
+ GENERATED SUMMARY:
47
+ South Africa's police are investigating the murder of a man who was shot in a pub in Soweto.
48
+
49
+ EVALUATION (fill in your analysis):
50
+ 1. Content Relevance (Does it capture main points?):
51
+ [Your evaluation: Score 1-5, explanation]
52
+
53
+ 2. Coherence (Does it flow well?):
54
+ [Your evaluation: Score 1-5, explanation]
55
+
56
+ 3. Fluency (Is it grammatically correct?):
57
+ [Your evaluation: Score 1-5, explanation]
58
+
59
+ 4. Factual Consistency (Are facts accurate?):
60
+ [Your evaluation: Score 1-5, explanation]
61
+
62
+ OVERALL SCORE: [X/20]
63
+
64
+ ================================================================================
65
+
66
+ ================================================================================
67
+ SAMPLE 3 (ROUGE-1: 0.923)
68
+ ================================================================================
69
+
70
+ DOCUMENT (first 300 characters):
71
+ The 27-year-old woman was found at Knowle Close, Redditch, after police were called early on Sunday.
72
+ The arrested man, a 32-year-old from Redditch, is in police custody.
73
+ The results of a post-mortem examination on the woman, who has not been formally identified, will be released later, a West Mercia...
74
+
75
+ REFERENCE SUMMARY:
76
+ A man has been arrested on suspicion of murder after a woman was found dead at a house in Worcestershire.
77
+
78
+ GENERATED SUMMARY:
79
+ A man has been arrested on suspicion of murder after a woman was found in a house in Worcestershire.
80
+
81
+ EVALUATION (fill in your analysis):
82
+ 1. Content Relevance (Does it capture main points?):
83
+ [Your evaluation: Score 1-5, explanation]
84
+
85
+ 2. Coherence (Does it flow well?):
86
+ [Your evaluation: Score 1-5, explanation]
87
+
88
+ 3. Fluency (Is it grammatically correct?):
89
+ [Your evaluation: Score 1-5, explanation]
90
+
91
+ 4. Factual Consistency (Are facts accurate?):
92
+ [Your evaluation: Score 1-5, explanation]
93
+
94
+ OVERALL SCORE: [X/20]
95
+
96
+ ================================================================================
MoE-Hash_metrics.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ROUGE-1": 29.39696373117908,
3
+ "ROUGE-2": 8.434329374459828,
4
+ "ROUGE-L": 22.575647505310812,
5
+ "BERTScore-P": 0.8836130499839783,
6
+ "BERTScore-R": 0.8762004971504211,
7
+ "BERTScore-F1": 0.8797832131385803,
8
+ "Compression Ratio": 0.08815832574259216,
9
+ "Extractiveness (%)": 53.66421487209081,
10
+ "Novel 1-grams (%)": 33.34628578269085,
11
+ "Novel 2-grams (%)": 74.40899167979951,
12
+ "Novel 3-grams (%)": 91.39474583114419,
13
+ "Novel 4-grams (%)": 96.61165474706667
14
+ }
MoE-Hash_predictions.txt ADDED
The diff for this file is too large to render. See raw diff
 
best_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e56659ee216a99e336ab6ba4f4cfe770468ed82ee58cb73d5852a1ee056126f
3
+ size 1235518587
config.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_dropout": 0.1,
3
+ "activation_function": "gelu",
4
+ "add_bias_logits": false,
5
+ "add_final_layer_norm": false,
6
+ "architectures": [
7
+ "BartForConditionalGenerationWithMoE"
8
+ ],
9
+ "attention_dropout": 0.1,
10
+ "bos_token_id": 0,
11
+ "classif_dropout": 0.1,
12
+ "classifier_dropout": 0.0,
13
+ "d_model": 768,
14
+ "decoder_attention_heads": 12,
15
+ "decoder_ffn_dim": 3072,
16
+ "decoder_layerdrop": 0.0,
17
+ "decoder_layers": 6,
18
+ "decoder_start_token_id": 2,
19
+ "dropout": 0.1,
20
+ "dtype": "float32",
21
+ "early_stopping": null,
22
+ "encoder_attention_heads": 12,
23
+ "encoder_ffn_dim": 3072,
24
+ "encoder_layerdrop": 0.0,
25
+ "encoder_layers": 6,
26
+ "eos_token_id": 2,
27
+ "forced_eos_token_id": 2,
28
+ "gradient_checkpointing": false,
29
+ "id2label": {
30
+ "0": "LABEL_0",
31
+ "1": "LABEL_1",
32
+ "2": "LABEL_2"
33
+ },
34
+ "init_std": 0.02,
35
+ "is_encoder_decoder": true,
36
+ "label2id": {
37
+ "LABEL_0": 0,
38
+ "LABEL_1": 1,
39
+ "LABEL_2": 2
40
+ },
41
+ "max_position_embeddings": 1024,
42
+ "model_type": "bart",
43
+ "no_repeat_ngram_size": null,
44
+ "normalize_before": false,
45
+ "normalize_embedding": true,
46
+ "num_beams": null,
47
+ "num_hidden_layers": 6,
48
+ "pad_token_id": 1,
49
+ "scale_embedding": false,
50
+ "task_specific_params": {
51
+ "summarization": {
52
+ "length_penalty": 1.0,
53
+ "max_length": 128,
54
+ "min_length": 12,
55
+ "num_beams": 4
56
+ },
57
+ "summarization_cnn": {
58
+ "length_penalty": 2.0,
59
+ "max_length": 142,
60
+ "min_length": 56,
61
+ "num_beams": 4
62
+ },
63
+ "summarization_xsum": {
64
+ "length_penalty": 1.0,
65
+ "max_length": 62,
66
+ "min_length": 11,
67
+ "num_beams": 6
68
+ }
69
+ },
70
+ "transformers_version": "4.57.1",
71
+ "use_cache": true,
72
+ "vocab_size": 50265
73
+ }
final_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4512d94c6d720436bb32519551f901b1253e0c5cdb1e273ce709b95f14da08b9
3
+ size 1235519690
generation_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 0,
4
+ "decoder_start_token_id": 2,
5
+ "early_stopping": true,
6
+ "eos_token_id": [
7
+ 2
8
+ ],
9
+ "forced_bos_token_id": 0,
10
+ "forced_eos_token_id": 2,
11
+ "no_repeat_ngram_size": 3,
12
+ "num_beams": 4,
13
+ "pad_token_id": 1,
14
+ "transformers_version": "4.57.1"
15
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19068bb1a1d6d1b03e2164b4b0a655bc498a26f3f61391d151ed68fff720b698
3
+ size 1464680804
moe_layers.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0075b4ecc85641e824626a648995756f5c3ae2c86c74a5159f7fccf90a14a5ab
3
+ size 906821341
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": true,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "<unk>",
46
+ "lstrip": false,
47
+ "normalized": true,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": true,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
+ "bos_token": "<s>",
46
+ "clean_up_tokenization_spaces": false,
47
+ "cls_token": "<s>",
48
+ "eos_token": "</s>",
49
+ "errors": "replace",
50
+ "extra_special_tokens": {},
51
+ "mask_token": "<mask>",
52
+ "model_max_length": 1000000000000000019884624838656,
53
+ "pad_token": "<pad>",
54
+ "sep_token": "</s>",
55
+ "tokenizer_class": "BartTokenizer",
56
+ "unk_token": "<unk>"
57
+ }
training_history.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "train_losses": [
3
+ 10.813822434134243,
4
+ 10.062575539820727,
5
+ 7.8534033919400805
6
+ ],
7
+ "val_losses": [
8
+ 10.822157826475824,
9
+ 9.285471734081609,
10
+ 6.486546616061097
11
+ ],
12
+ "best_val_loss": 6.486546616061097
13
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff