master-mahdi commited on
Commit
b659dd0
·
verified ·
1 Parent(s): 58abf9a

master-mahdi/persian_intent_detection

Browse files
README.md ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: HooshvareLab/bert-fa-base-uncased-clf-persiannews
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
+ model-index:
9
+ - name: my_model
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # my_model
17
+
18
+ This model is a fine-tuned version of [HooshvareLab/bert-fa-base-uncased-clf-persiannews](https://huggingface.co/HooshvareLab/bert-fa-base-uncased-clf-persiannews) on an unknown dataset.
19
+ It achieves the following results on the evaluation set:
20
+ - Loss: 0.8573
21
+ - Accuracy: 0.8695
22
+
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 2e-05
41
+ - train_batch_size: 32
42
+ - eval_batch_size: 32
43
+ - seed: 42
44
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
+ - lr_scheduler_type: linear
46
+ - num_epochs: 20
47
+
48
+ ### Training results
49
+
50
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
51
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|
52
+ | No log | 1.0 | 360 | 1.0779 | 0.7772 |
53
+ | 1.8476 | 2.0 | 720 | 0.7038 | 0.8436 |
54
+ | 0.6038 | 3.0 | 1080 | 0.6356 | 0.8490 |
55
+ | 0.6038 | 4.0 | 1440 | 0.6477 | 0.8470 |
56
+ | 0.3037 | 5.0 | 1800 | 0.6521 | 0.8515 |
57
+ | 0.1723 | 6.0 | 2160 | 0.6938 | 0.8554 |
58
+ | 0.1066 | 7.0 | 2520 | 0.6942 | 0.8519 |
59
+ | 0.1066 | 8.0 | 2880 | 0.7426 | 0.8559 |
60
+ | 0.0636 | 9.0 | 3240 | 0.7769 | 0.8549 |
61
+ | 0.0464 | 10.0 | 3600 | 0.8179 | 0.8549 |
62
+ | 0.0464 | 11.0 | 3960 | 0.8457 | 0.8544 |
63
+ | 0.029 | 12.0 | 4320 | 0.8509 | 0.8554 |
64
+ | 0.0241 | 13.0 | 4680 | 0.8534 | 0.8529 |
65
+ | 0.0158 | 14.0 | 5040 | 0.8793 | 0.8549 |
66
+ | 0.0158 | 15.0 | 5400 | 0.9053 | 0.8544 |
67
+ | 0.0131 | 16.0 | 5760 | 0.9009 | 0.8539 |
68
+ | 0.012 | 17.0 | 6120 | 0.9124 | 0.8554 |
69
+ | 0.012 | 18.0 | 6480 | 0.9092 | 0.8549 |
70
+ | 0.0114 | 19.0 | 6840 | 0.9117 | 0.8529 |
71
+ | 0.0079 | 20.0 | 7200 | 0.9129 | 0.8519 |
72
+
73
+
74
+ ### Framework versions
75
+
76
+ - Transformers 4.41.2
77
+ - Pytorch 2.3.0+cu121
78
+ - Datasets 2.20.0
79
+ - Tokenizers 0.19.1
config.json ADDED
@@ -0,0 +1,152 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "HooshvareLab/bert-fa-base-uncased-clf-persiannews",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "finetuning_task": "persiannews",
9
+ "gradient_checkpointing": false,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
+ "id2label": {
14
+ "0": "datetime_query",
15
+ "1": "iot_hue_lightchange",
16
+ "2": "transport_ticket",
17
+ "3": "takeaway_query",
18
+ "4": "qa_stock",
19
+ "5": "general_greet",
20
+ "6": "recommendation_events",
21
+ "7": "music_dislikeness",
22
+ "8": "iot_wemo_off",
23
+ "9": "cooking_recipe",
24
+ "10": "qa_currency",
25
+ "11": "transport_traffic",
26
+ "12": "general_quirky",
27
+ "13": "weather_query",
28
+ "14": "audio_volume_up",
29
+ "15": "email_addcontact",
30
+ "16": "takeaway_order",
31
+ "17": "email_querycontact",
32
+ "18": "iot_hue_lightup",
33
+ "19": "recommendation_locations",
34
+ "20": "play_audiobook",
35
+ "21": "lists_createoradd",
36
+ "22": "news_query",
37
+ "23": "alarm_query",
38
+ "24": "iot_wemo_on",
39
+ "25": "general_joke",
40
+ "26": "qa_definition",
41
+ "27": "social_query",
42
+ "28": "music_settings",
43
+ "29": "audio_volume_other",
44
+ "30": "calendar_remove",
45
+ "31": "iot_hue_lightdim",
46
+ "32": "calendar_query",
47
+ "33": "email_sendemail",
48
+ "34": "iot_cleaning",
49
+ "35": "audio_volume_down",
50
+ "36": "play_radio",
51
+ "37": "cooking_query",
52
+ "38": "datetime_convert",
53
+ "39": "qa_maths",
54
+ "40": "iot_hue_lightoff",
55
+ "41": "iot_hue_lighton",
56
+ "42": "transport_query",
57
+ "43": "music_likeness",
58
+ "44": "email_query",
59
+ "45": "play_music",
60
+ "46": "audio_volume_mute",
61
+ "47": "social_post",
62
+ "48": "alarm_set",
63
+ "49": "qa_factoid",
64
+ "50": "calendar_set",
65
+ "51": "play_game",
66
+ "52": "alarm_remove",
67
+ "53": "lists_remove",
68
+ "54": "transport_taxi",
69
+ "55": "recommendation_movies",
70
+ "56": "iot_coffee",
71
+ "57": "music_query",
72
+ "58": "play_podcasts",
73
+ "59": "lists_query"
74
+ },
75
+ "initializer_range": 0.02,
76
+ "intermediate_size": 3072,
77
+ "label2id": {
78
+ "alarm_query": 23,
79
+ "alarm_remove": 52,
80
+ "alarm_set": 48,
81
+ "audio_volume_down": 35,
82
+ "audio_volume_mute": 46,
83
+ "audio_volume_other": 29,
84
+ "audio_volume_up": 14,
85
+ "calendar_query": 32,
86
+ "calendar_remove": 30,
87
+ "calendar_set": 50,
88
+ "cooking_query": 37,
89
+ "cooking_recipe": 9,
90
+ "datetime_convert": 38,
91
+ "datetime_query": 0,
92
+ "email_addcontact": 15,
93
+ "email_query": 44,
94
+ "email_querycontact": 17,
95
+ "email_sendemail": 33,
96
+ "general_greet": 5,
97
+ "general_joke": 25,
98
+ "general_quirky": 12,
99
+ "iot_cleaning": 34,
100
+ "iot_coffee": 56,
101
+ "iot_hue_lightchange": 1,
102
+ "iot_hue_lightdim": 31,
103
+ "iot_hue_lightoff": 40,
104
+ "iot_hue_lighton": 41,
105
+ "iot_hue_lightup": 18,
106
+ "iot_wemo_off": 8,
107
+ "iot_wemo_on": 24,
108
+ "lists_createoradd": 21,
109
+ "lists_query": 59,
110
+ "lists_remove": 53,
111
+ "music_dislikeness": 7,
112
+ "music_likeness": 43,
113
+ "music_query": 57,
114
+ "music_settings": 28,
115
+ "news_query": 22,
116
+ "play_audiobook": 20,
117
+ "play_game": 51,
118
+ "play_music": 45,
119
+ "play_podcasts": 58,
120
+ "play_radio": 36,
121
+ "qa_currency": 10,
122
+ "qa_definition": 26,
123
+ "qa_factoid": 49,
124
+ "qa_maths": 39,
125
+ "qa_stock": 4,
126
+ "recommendation_events": 6,
127
+ "recommendation_locations": 19,
128
+ "recommendation_movies": 55,
129
+ "social_post": 47,
130
+ "social_query": 27,
131
+ "takeaway_order": 16,
132
+ "takeaway_query": 3,
133
+ "transport_query": 42,
134
+ "transport_taxi": 54,
135
+ "transport_ticket": 2,
136
+ "transport_traffic": 11,
137
+ "weather_query": 13
138
+ },
139
+ "layer_norm_eps": 1e-12,
140
+ "max_position_embeddings": 512,
141
+ "model_type": "bert",
142
+ "num_attention_heads": 12,
143
+ "num_hidden_layers": 12,
144
+ "pad_token_id": 0,
145
+ "position_embedding_type": "absolute",
146
+ "problem_type": "single_label_classification",
147
+ "torch_dtype": "float32",
148
+ "transformers_version": "4.41.2",
149
+ "type_vocab_size": 2,
150
+ "use_cache": true,
151
+ "vocab_size": 100000
152
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b911302678e5e6dd525225ad623ac901d6f03fed4a6129311332d38288a14fe
3
+ size 651573488
runs/Jun21_12-18-44_175182b8c458/events.out.tfevents.1718972325.175182b8c458.519.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1476d741d7e126c184a86cd32c0ad717114d89d2d1686694cd4d3e5e8e48568
3
+ size 17861
runs/Jun21_12-18-44_175182b8c458/events.out.tfevents.1718974108.175182b8c458.519.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4a8a1e067708190b1c8b0ebf6188772ff4e9a776d776e2dbe7ed9b9cd57dfd2
3
+ size 411
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[MASK]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "4": {
36
+ "content": "[SEP]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "full_tokenizer_file": null,
49
+ "mask_token": "[MASK]",
50
+ "model_max_length": 1000000000000000019884624838656,
51
+ "never_split": null,
52
+ "pad_token": "[PAD]",
53
+ "sep_token": "[SEP]",
54
+ "strip_accents": null,
55
+ "tokenize_chinese_chars": true,
56
+ "tokenizer_class": "BertTokenizer",
57
+ "unk_token": "[UNK]"
58
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2fda16c315584deeac9425f075dd91098ee4651864ae11ece96f67600533123
3
+ size 5112
vocab.txt ADDED
The diff for this file is too large to render. See raw diff