PuxAI commited on
Commit
1bb796a
·
verified ·
1 Parent(s): bcf56bc

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/config.json +155 -0
  2. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/model.safetensors +3 -0
  3. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/optimizer.pt +3 -0
  4. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/rng_state.pth +3 -0
  5. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/scheduler.pt +3 -0
  6. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/tokenizer.json +0 -0
  7. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/tokenizer_config.json +23 -0
  8. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/trainer_state.json +48 -0
  9. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/training_args.bin +3 -0
  10. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/config.json +155 -0
  11. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/model.safetensors +3 -0
  12. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/optimizer.pt +3 -0
  13. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/rng_state.pth +3 -0
  14. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/scheduler.pt +3 -0
  15. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/tokenizer.json +0 -0
  16. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/tokenizer_config.json +23 -0
  17. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/trainer_state.json +174 -0
  18. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/training_args.bin +3 -0
  19. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/config.json +155 -0
  20. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/model.safetensors +3 -0
  21. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/optimizer.pt +3 -0
  22. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/rng_state.pth +3 -0
  23. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/scheduler.pt +3 -0
  24. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/tokenizer.json +0 -0
  25. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/tokenizer_config.json +23 -0
  26. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/trainer_state.json +181 -0
  27. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/training_args.bin +3 -0
  28. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/config.json +155 -0
  29. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/model.safetensors +3 -0
  30. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/optimizer.pt +3 -0
  31. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/rng_state.pth +3 -0
  32. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/scheduler.pt +3 -0
  33. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/tokenizer.json +0 -0
  34. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/tokenizer_config.json +23 -0
  35. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/trainer_state.json +188 -0
  36. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/training_args.bin +3 -0
  37. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/config.json +155 -0
  38. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/model.safetensors +3 -0
  39. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/optimizer.pt +3 -0
  40. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/rng_state.pth +3 -0
  41. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/scheduler.pt +3 -0
  42. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/tokenizer.json +0 -0
  43. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/tokenizer_config.json +23 -0
  44. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/trainer_state.json +195 -0
  45. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/training_args.bin +3 -0
  46. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-12000/config.json +155 -0
  47. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-12000/model.safetensors +3 -0
  48. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-12000/optimizer.pt +3 -0
  49. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-12000/rng_state.pth +3 -0
  50. nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-12000/scheduler.pt +3 -0
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/config.json ADDED
@@ -0,0 +1,155 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "DebertaAdvancedSpanClassifier"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": null,
7
+ "dtype": "float32",
8
+ "eos_token_id": null,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "O",
14
+ "1": "ACCOUNT_NUMBER",
15
+ "2": "AGE",
16
+ "3": "API_KEY",
17
+ "4": "BANK_ROUTING_NUMBER",
18
+ "5": "BIOMETRIC_IDENTIFIER",
19
+ "6": "BLOOD_TYPE",
20
+ "7": "CERTIFICATE_LICENSE_NUMBER",
21
+ "8": "CITY",
22
+ "9": "COMPANY_NAME",
23
+ "10": "COORDINATE",
24
+ "11": "COUNTRY",
25
+ "12": "COUNTY",
26
+ "13": "CREDIT_DEBIT_CARD",
27
+ "14": "CUSTOMER_ID",
28
+ "15": "CVV",
29
+ "16": "DATE",
30
+ "17": "DATE_OF_BIRTH",
31
+ "18": "DATE_TIME",
32
+ "19": "DEVICE_IDENTIFIER",
33
+ "20": "EDUCATION_LEVEL",
34
+ "21": "EMAIL",
35
+ "22": "EMPLOYEE_ID",
36
+ "23": "EMPLOYMENT_STATUS",
37
+ "24": "FAX_NUMBER",
38
+ "25": "FIRST_NAME",
39
+ "26": "GENDER",
40
+ "27": "HEALTH_PLAN_BENEFICIARY_NUMBER",
41
+ "28": "HTTP_COOKIE",
42
+ "29": "IPV4",
43
+ "30": "IPV6",
44
+ "31": "LANGUAGE",
45
+ "32": "LAST_NAME",
46
+ "33": "LICENSE_PLATE",
47
+ "34": "MAC_ADDRESS",
48
+ "35": "MEDICAL_RECORD_NUMBER",
49
+ "36": "NATIONAL_ID",
50
+ "37": "OCCUPATION",
51
+ "38": "PASSWORD",
52
+ "39": "PHONE_NUMBER",
53
+ "40": "PIN",
54
+ "41": "POLITICAL_VIEW",
55
+ "42": "POSTCODE",
56
+ "43": "RACE_ETHNICITY",
57
+ "44": "RELIGIOUS_BELIEF",
58
+ "45": "SEXUALITY",
59
+ "46": "SSN",
60
+ "47": "STATE",
61
+ "48": "STREET_ADDRESS",
62
+ "49": "SWIFT_BIC",
63
+ "50": "TAX_ID",
64
+ "51": "TIME",
65
+ "52": "UNIQUE_ID",
66
+ "53": "URL",
67
+ "54": "USER_NAME",
68
+ "55": "VEHICLE_IDENTIFIER"
69
+ },
70
+ "initializer_range": 0.02,
71
+ "intermediate_size": 3072,
72
+ "label2id": {
73
+ "ACCOUNT_NUMBER": 1,
74
+ "AGE": 2,
75
+ "API_KEY": 3,
76
+ "BANK_ROUTING_NUMBER": 4,
77
+ "BIOMETRIC_IDENTIFIER": 5,
78
+ "BLOOD_TYPE": 6,
79
+ "CERTIFICATE_LICENSE_NUMBER": 7,
80
+ "CITY": 8,
81
+ "COMPANY_NAME": 9,
82
+ "COORDINATE": 10,
83
+ "COUNTRY": 11,
84
+ "COUNTY": 12,
85
+ "CREDIT_DEBIT_CARD": 13,
86
+ "CUSTOMER_ID": 14,
87
+ "CVV": 15,
88
+ "DATE": 16,
89
+ "DATE_OF_BIRTH": 17,
90
+ "DATE_TIME": 18,
91
+ "DEVICE_IDENTIFIER": 19,
92
+ "EDUCATION_LEVEL": 20,
93
+ "EMAIL": 21,
94
+ "EMPLOYEE_ID": 22,
95
+ "EMPLOYMENT_STATUS": 23,
96
+ "FAX_NUMBER": 24,
97
+ "FIRST_NAME": 25,
98
+ "GENDER": 26,
99
+ "HEALTH_PLAN_BENEFICIARY_NUMBER": 27,
100
+ "HTTP_COOKIE": 28,
101
+ "IPV4": 29,
102
+ "IPV6": 30,
103
+ "LANGUAGE": 31,
104
+ "LAST_NAME": 32,
105
+ "LICENSE_PLATE": 33,
106
+ "MAC_ADDRESS": 34,
107
+ "MEDICAL_RECORD_NUMBER": 35,
108
+ "NATIONAL_ID": 36,
109
+ "O": 0,
110
+ "OCCUPATION": 37,
111
+ "PASSWORD": 38,
112
+ "PHONE_NUMBER": 39,
113
+ "PIN": 40,
114
+ "POLITICAL_VIEW": 41,
115
+ "POSTCODE": 42,
116
+ "RACE_ETHNICITY": 43,
117
+ "RELIGIOUS_BELIEF": 44,
118
+ "SEXUALITY": 45,
119
+ "SSN": 46,
120
+ "STATE": 47,
121
+ "STREET_ADDRESS": 48,
122
+ "SWIFT_BIC": 49,
123
+ "TAX_ID": 50,
124
+ "TIME": 51,
125
+ "UNIQUE_ID": 52,
126
+ "URL": 53,
127
+ "USER_NAME": 54,
128
+ "VEHICLE_IDENTIFIER": 55
129
+ },
130
+ "layer_norm_eps": 1e-07,
131
+ "legacy": true,
132
+ "max_position_embeddings": 512,
133
+ "max_relative_positions": -1,
134
+ "model_type": "deberta-v2",
135
+ "norm_rel_ebd": "layer_norm",
136
+ "num_attention_heads": 12,
137
+ "num_hidden_layers": 12,
138
+ "pad_token_id": 0,
139
+ "pooler_dropout": 0,
140
+ "pooler_hidden_act": "gelu",
141
+ "pooler_hidden_size": 768,
142
+ "pos_att_type": [
143
+ "p2c",
144
+ "c2p"
145
+ ],
146
+ "position_biased_input": false,
147
+ "position_buckets": 256,
148
+ "relative_attention": true,
149
+ "share_att_key": true,
150
+ "tie_word_embeddings": true,
151
+ "transformers_version": "5.3.0",
152
+ "type_vocab_size": 0,
153
+ "use_cache": false,
154
+ "vocab_size": 128100
155
+ }
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10683c7e1c5045afaa3aeacc26936db9e7c0c1816ff8acccd56a87950f59b656
3
+ size 741501240
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1d4f3787cda2a35c3597e3afddc54b76cb8bac926dd34efe503f02a0c7f45d7
3
+ size 1483126155
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0898f1c98660ddce6c8fa2b024604b68897e1de8f58587a57075b17149209488
3
+ size 14645
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5fea4e0656adf3f60d71d6abc0a4fd2213c272c07b9cc45a432cd7ce0cdc5f1
3
+ size 1465
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/tokenizer_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "backend": "tokenizers",
4
+ "bos_token": "[CLS]",
5
+ "cls_token": "[CLS]",
6
+ "do_lower_case": false,
7
+ "eos_token": "[SEP]",
8
+ "extra_special_tokens": [
9
+ "[PAD]",
10
+ "[CLS]",
11
+ "[SEP]"
12
+ ],
13
+ "is_local": false,
14
+ "mask_token": "[MASK]",
15
+ "model_max_length": 1000000000000000019884624838656,
16
+ "pad_token": "[PAD]",
17
+ "sep_token": "[SEP]",
18
+ "split_by_punct": false,
19
+ "tokenizer_class": "DebertaV2Tokenizer",
20
+ "unk_id": 3,
21
+ "unk_token": "[UNK]",
22
+ "vocab_type": "spm"
23
+ }
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/trainer_state.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.16,
6
+ "eval_steps": 500,
7
+ "global_step": 1000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.08,
14
+ "grad_norm": 0.003990198019891977,
15
+ "learning_rate": 1.9201600000000003e-05,
16
+ "loss": 0.005083163738250732,
17
+ "step": 500
18
+ },
19
+ {
20
+ "epoch": 0.16,
21
+ "grad_norm": 0.006030593533068895,
22
+ "learning_rate": 1.84016e-05,
23
+ "loss": 0.0017981358766555786,
24
+ "step": 1000
25
+ }
26
+ ],
27
+ "logging_steps": 500,
28
+ "max_steps": 12500,
29
+ "num_input_tokens_seen": 0,
30
+ "num_train_epochs": 2,
31
+ "save_steps": 500,
32
+ "stateful_callbacks": {
33
+ "TrainerControl": {
34
+ "args": {
35
+ "should_epoch_stop": false,
36
+ "should_evaluate": false,
37
+ "should_log": false,
38
+ "should_save": true,
39
+ "should_training_stop": false
40
+ },
41
+ "attributes": {}
42
+ }
43
+ },
44
+ "total_flos": 4058261169755904.0,
45
+ "train_batch_size": 16,
46
+ "trial_name": null,
47
+ "trial_params": null
48
+ }
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-1000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8038413655afcd9f5ce084a60e4a073f6b35472739c0d53f5df4cbdca825d93e
3
+ size 5201
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/config.json ADDED
@@ -0,0 +1,155 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "DebertaAdvancedSpanClassifier"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": null,
7
+ "dtype": "float32",
8
+ "eos_token_id": null,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "O",
14
+ "1": "ACCOUNT_NUMBER",
15
+ "2": "AGE",
16
+ "3": "API_KEY",
17
+ "4": "BANK_ROUTING_NUMBER",
18
+ "5": "BIOMETRIC_IDENTIFIER",
19
+ "6": "BLOOD_TYPE",
20
+ "7": "CERTIFICATE_LICENSE_NUMBER",
21
+ "8": "CITY",
22
+ "9": "COMPANY_NAME",
23
+ "10": "COORDINATE",
24
+ "11": "COUNTRY",
25
+ "12": "COUNTY",
26
+ "13": "CREDIT_DEBIT_CARD",
27
+ "14": "CUSTOMER_ID",
28
+ "15": "CVV",
29
+ "16": "DATE",
30
+ "17": "DATE_OF_BIRTH",
31
+ "18": "DATE_TIME",
32
+ "19": "DEVICE_IDENTIFIER",
33
+ "20": "EDUCATION_LEVEL",
34
+ "21": "EMAIL",
35
+ "22": "EMPLOYEE_ID",
36
+ "23": "EMPLOYMENT_STATUS",
37
+ "24": "FAX_NUMBER",
38
+ "25": "FIRST_NAME",
39
+ "26": "GENDER",
40
+ "27": "HEALTH_PLAN_BENEFICIARY_NUMBER",
41
+ "28": "HTTP_COOKIE",
42
+ "29": "IPV4",
43
+ "30": "IPV6",
44
+ "31": "LANGUAGE",
45
+ "32": "LAST_NAME",
46
+ "33": "LICENSE_PLATE",
47
+ "34": "MAC_ADDRESS",
48
+ "35": "MEDICAL_RECORD_NUMBER",
49
+ "36": "NATIONAL_ID",
50
+ "37": "OCCUPATION",
51
+ "38": "PASSWORD",
52
+ "39": "PHONE_NUMBER",
53
+ "40": "PIN",
54
+ "41": "POLITICAL_VIEW",
55
+ "42": "POSTCODE",
56
+ "43": "RACE_ETHNICITY",
57
+ "44": "RELIGIOUS_BELIEF",
58
+ "45": "SEXUALITY",
59
+ "46": "SSN",
60
+ "47": "STATE",
61
+ "48": "STREET_ADDRESS",
62
+ "49": "SWIFT_BIC",
63
+ "50": "TAX_ID",
64
+ "51": "TIME",
65
+ "52": "UNIQUE_ID",
66
+ "53": "URL",
67
+ "54": "USER_NAME",
68
+ "55": "VEHICLE_IDENTIFIER"
69
+ },
70
+ "initializer_range": 0.02,
71
+ "intermediate_size": 3072,
72
+ "label2id": {
73
+ "ACCOUNT_NUMBER": 1,
74
+ "AGE": 2,
75
+ "API_KEY": 3,
76
+ "BANK_ROUTING_NUMBER": 4,
77
+ "BIOMETRIC_IDENTIFIER": 5,
78
+ "BLOOD_TYPE": 6,
79
+ "CERTIFICATE_LICENSE_NUMBER": 7,
80
+ "CITY": 8,
81
+ "COMPANY_NAME": 9,
82
+ "COORDINATE": 10,
83
+ "COUNTRY": 11,
84
+ "COUNTY": 12,
85
+ "CREDIT_DEBIT_CARD": 13,
86
+ "CUSTOMER_ID": 14,
87
+ "CVV": 15,
88
+ "DATE": 16,
89
+ "DATE_OF_BIRTH": 17,
90
+ "DATE_TIME": 18,
91
+ "DEVICE_IDENTIFIER": 19,
92
+ "EDUCATION_LEVEL": 20,
93
+ "EMAIL": 21,
94
+ "EMPLOYEE_ID": 22,
95
+ "EMPLOYMENT_STATUS": 23,
96
+ "FAX_NUMBER": 24,
97
+ "FIRST_NAME": 25,
98
+ "GENDER": 26,
99
+ "HEALTH_PLAN_BENEFICIARY_NUMBER": 27,
100
+ "HTTP_COOKIE": 28,
101
+ "IPV4": 29,
102
+ "IPV6": 30,
103
+ "LANGUAGE": 31,
104
+ "LAST_NAME": 32,
105
+ "LICENSE_PLATE": 33,
106
+ "MAC_ADDRESS": 34,
107
+ "MEDICAL_RECORD_NUMBER": 35,
108
+ "NATIONAL_ID": 36,
109
+ "O": 0,
110
+ "OCCUPATION": 37,
111
+ "PASSWORD": 38,
112
+ "PHONE_NUMBER": 39,
113
+ "PIN": 40,
114
+ "POLITICAL_VIEW": 41,
115
+ "POSTCODE": 42,
116
+ "RACE_ETHNICITY": 43,
117
+ "RELIGIOUS_BELIEF": 44,
118
+ "SEXUALITY": 45,
119
+ "SSN": 46,
120
+ "STATE": 47,
121
+ "STREET_ADDRESS": 48,
122
+ "SWIFT_BIC": 49,
123
+ "TAX_ID": 50,
124
+ "TIME": 51,
125
+ "UNIQUE_ID": 52,
126
+ "URL": 53,
127
+ "USER_NAME": 54,
128
+ "VEHICLE_IDENTIFIER": 55
129
+ },
130
+ "layer_norm_eps": 1e-07,
131
+ "legacy": true,
132
+ "max_position_embeddings": 512,
133
+ "max_relative_positions": -1,
134
+ "model_type": "deberta-v2",
135
+ "norm_rel_ebd": "layer_norm",
136
+ "num_attention_heads": 12,
137
+ "num_hidden_layers": 12,
138
+ "pad_token_id": 0,
139
+ "pooler_dropout": 0,
140
+ "pooler_hidden_act": "gelu",
141
+ "pooler_hidden_size": 768,
142
+ "pos_att_type": [
143
+ "p2c",
144
+ "c2p"
145
+ ],
146
+ "position_biased_input": false,
147
+ "position_buckets": 256,
148
+ "relative_attention": true,
149
+ "share_att_key": true,
150
+ "tie_word_embeddings": true,
151
+ "transformers_version": "5.3.0",
152
+ "type_vocab_size": 0,
153
+ "use_cache": false,
154
+ "vocab_size": 128100
155
+ }
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdc7ed18a6cf403a7fd2f0f5110c3575c76c1c8f4ad4687a60e94b5c0d6b6279
3
+ size 741501240
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5984dc3dc9885a3dc13068304f187bd65d398638ddbfeededca242e259735055
3
+ size 1483126155
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa8cdbee539d97bae70366a03261561536d13727025d3799e8817cde8a1db61e
3
+ size 14645
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b5b98e0636e3f3c3e0e2b08a05e847412f02edd7b453ed06b824475431489c1
3
+ size 1465
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/tokenizer_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "backend": "tokenizers",
4
+ "bos_token": "[CLS]",
5
+ "cls_token": "[CLS]",
6
+ "do_lower_case": false,
7
+ "eos_token": "[SEP]",
8
+ "extra_special_tokens": [
9
+ "[PAD]",
10
+ "[CLS]",
11
+ "[SEP]"
12
+ ],
13
+ "is_local": false,
14
+ "mask_token": "[MASK]",
15
+ "model_max_length": 1000000000000000019884624838656,
16
+ "pad_token": "[PAD]",
17
+ "sep_token": "[SEP]",
18
+ "split_by_punct": false,
19
+ "tokenizer_class": "DebertaV2Tokenizer",
20
+ "unk_id": 3,
21
+ "unk_token": "[UNK]",
22
+ "vocab_type": "spm"
23
+ }
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/trainer_state.json ADDED
@@ -0,0 +1,174 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.6,
6
+ "eval_steps": 500,
7
+ "global_step": 10000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.08,
14
+ "grad_norm": 0.003990198019891977,
15
+ "learning_rate": 1.9201600000000003e-05,
16
+ "loss": 0.005083163738250732,
17
+ "step": 500
18
+ },
19
+ {
20
+ "epoch": 0.16,
21
+ "grad_norm": 0.006030593533068895,
22
+ "learning_rate": 1.84016e-05,
23
+ "loss": 0.0017981358766555786,
24
+ "step": 1000
25
+ },
26
+ {
27
+ "epoch": 0.24,
28
+ "grad_norm": 0.0030848984606564045,
29
+ "learning_rate": 1.76016e-05,
30
+ "loss": 0.0008149439096450806,
31
+ "step": 1500
32
+ },
33
+ {
34
+ "epoch": 0.32,
35
+ "grad_norm": 0.002702153753489256,
36
+ "learning_rate": 1.6801600000000002e-05,
37
+ "loss": 0.0004432150423526764,
38
+ "step": 2000
39
+ },
40
+ {
41
+ "epoch": 0.4,
42
+ "grad_norm": 0.002467973157763481,
43
+ "learning_rate": 1.6001600000000003e-05,
44
+ "loss": 0.00028532958030700684,
45
+ "step": 2500
46
+ },
47
+ {
48
+ "epoch": 0.48,
49
+ "grad_norm": 0.0021964123006910086,
50
+ "learning_rate": 1.52016e-05,
51
+ "loss": 0.00020916727185249327,
52
+ "step": 3000
53
+ },
54
+ {
55
+ "epoch": 0.56,
56
+ "grad_norm": 0.004308697767555714,
57
+ "learning_rate": 1.4401600000000003e-05,
58
+ "loss": 0.00017732538282871246,
59
+ "step": 3500
60
+ },
61
+ {
62
+ "epoch": 0.64,
63
+ "grad_norm": 0.002398895798251033,
64
+ "learning_rate": 1.3601600000000002e-05,
65
+ "loss": 0.000154940202832222,
66
+ "step": 4000
67
+ },
68
+ {
69
+ "epoch": 0.72,
70
+ "grad_norm": 0.0022735886741429567,
71
+ "learning_rate": 1.2801600000000001e-05,
72
+ "loss": 0.00014079591631889344,
73
+ "step": 4500
74
+ },
75
+ {
76
+ "epoch": 0.8,
77
+ "grad_norm": 0.00556814344599843,
78
+ "learning_rate": 1.20016e-05,
79
+ "loss": 0.0001373387724161148,
80
+ "step": 5000
81
+ },
82
+ {
83
+ "epoch": 0.88,
84
+ "grad_norm": 0.0012896452099084854,
85
+ "learning_rate": 1.1201600000000001e-05,
86
+ "loss": 0.00012117443978786468,
87
+ "step": 5500
88
+ },
89
+ {
90
+ "epoch": 0.96,
91
+ "grad_norm": 0.0033518935088068247,
92
+ "learning_rate": 1.04016e-05,
93
+ "loss": 0.00011866840720176697,
94
+ "step": 6000
95
+ },
96
+ {
97
+ "epoch": 1.04,
98
+ "grad_norm": 0.0012641055509448051,
99
+ "learning_rate": 9.601600000000001e-06,
100
+ "loss": 0.0001060236245393753,
101
+ "step": 6500
102
+ },
103
+ {
104
+ "epoch": 1.12,
105
+ "grad_norm": 0.000537463347427547,
106
+ "learning_rate": 8.8016e-06,
107
+ "loss": 0.00010027290880680084,
108
+ "step": 7000
109
+ },
110
+ {
111
+ "epoch": 1.2,
112
+ "grad_norm": 0.0018809024477377534,
113
+ "learning_rate": 8.0016e-06,
114
+ "loss": 9.29199755191803e-05,
115
+ "step": 7500
116
+ },
117
+ {
118
+ "epoch": 1.28,
119
+ "grad_norm": 0.005723062437027693,
120
+ "learning_rate": 7.201600000000001e-06,
121
+ "loss": 9.09605547785759e-05,
122
+ "step": 8000
123
+ },
124
+ {
125
+ "epoch": 1.3599999999999999,
126
+ "grad_norm": 0.0005352628650143743,
127
+ "learning_rate": 6.4016e-06,
128
+ "loss": 9.268201887607574e-05,
129
+ "step": 8500
130
+ },
131
+ {
132
+ "epoch": 1.44,
133
+ "grad_norm": 0.001981391105800867,
134
+ "learning_rate": 5.601600000000001e-06,
135
+ "loss": 8.78596231341362e-05,
136
+ "step": 9000
137
+ },
138
+ {
139
+ "epoch": 1.52,
140
+ "grad_norm": 0.007651986088603735,
141
+ "learning_rate": 4.8016e-06,
142
+ "loss": 8.183776587247848e-05,
143
+ "step": 9500
144
+ },
145
+ {
146
+ "epoch": 1.6,
147
+ "grad_norm": 0.002296850783750415,
148
+ "learning_rate": 4.001600000000001e-06,
149
+ "loss": 8.76401886343956e-05,
150
+ "step": 10000
151
+ }
152
+ ],
153
+ "logging_steps": 500,
154
+ "max_steps": 12500,
155
+ "num_input_tokens_seen": 0,
156
+ "num_train_epochs": 2,
157
+ "save_steps": 500,
158
+ "stateful_callbacks": {
159
+ "TrainerControl": {
160
+ "args": {
161
+ "should_epoch_stop": false,
162
+ "should_evaluate": false,
163
+ "should_log": false,
164
+ "should_save": true,
165
+ "should_training_stop": false
166
+ },
167
+ "attributes": {}
168
+ }
169
+ },
170
+ "total_flos": 4.077091226313677e+16,
171
+ "train_batch_size": 16,
172
+ "trial_name": null,
173
+ "trial_params": null
174
+ }
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8038413655afcd9f5ce084a60e4a073f6b35472739c0d53f5df4cbdca825d93e
3
+ size 5201
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/config.json ADDED
@@ -0,0 +1,155 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "DebertaAdvancedSpanClassifier"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": null,
7
+ "dtype": "float32",
8
+ "eos_token_id": null,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "O",
14
+ "1": "ACCOUNT_NUMBER",
15
+ "2": "AGE",
16
+ "3": "API_KEY",
17
+ "4": "BANK_ROUTING_NUMBER",
18
+ "5": "BIOMETRIC_IDENTIFIER",
19
+ "6": "BLOOD_TYPE",
20
+ "7": "CERTIFICATE_LICENSE_NUMBER",
21
+ "8": "CITY",
22
+ "9": "COMPANY_NAME",
23
+ "10": "COORDINATE",
24
+ "11": "COUNTRY",
25
+ "12": "COUNTY",
26
+ "13": "CREDIT_DEBIT_CARD",
27
+ "14": "CUSTOMER_ID",
28
+ "15": "CVV",
29
+ "16": "DATE",
30
+ "17": "DATE_OF_BIRTH",
31
+ "18": "DATE_TIME",
32
+ "19": "DEVICE_IDENTIFIER",
33
+ "20": "EDUCATION_LEVEL",
34
+ "21": "EMAIL",
35
+ "22": "EMPLOYEE_ID",
36
+ "23": "EMPLOYMENT_STATUS",
37
+ "24": "FAX_NUMBER",
38
+ "25": "FIRST_NAME",
39
+ "26": "GENDER",
40
+ "27": "HEALTH_PLAN_BENEFICIARY_NUMBER",
41
+ "28": "HTTP_COOKIE",
42
+ "29": "IPV4",
43
+ "30": "IPV6",
44
+ "31": "LANGUAGE",
45
+ "32": "LAST_NAME",
46
+ "33": "LICENSE_PLATE",
47
+ "34": "MAC_ADDRESS",
48
+ "35": "MEDICAL_RECORD_NUMBER",
49
+ "36": "NATIONAL_ID",
50
+ "37": "OCCUPATION",
51
+ "38": "PASSWORD",
52
+ "39": "PHONE_NUMBER",
53
+ "40": "PIN",
54
+ "41": "POLITICAL_VIEW",
55
+ "42": "POSTCODE",
56
+ "43": "RACE_ETHNICITY",
57
+ "44": "RELIGIOUS_BELIEF",
58
+ "45": "SEXUALITY",
59
+ "46": "SSN",
60
+ "47": "STATE",
61
+ "48": "STREET_ADDRESS",
62
+ "49": "SWIFT_BIC",
63
+ "50": "TAX_ID",
64
+ "51": "TIME",
65
+ "52": "UNIQUE_ID",
66
+ "53": "URL",
67
+ "54": "USER_NAME",
68
+ "55": "VEHICLE_IDENTIFIER"
69
+ },
70
+ "initializer_range": 0.02,
71
+ "intermediate_size": 3072,
72
+ "label2id": {
73
+ "ACCOUNT_NUMBER": 1,
74
+ "AGE": 2,
75
+ "API_KEY": 3,
76
+ "BANK_ROUTING_NUMBER": 4,
77
+ "BIOMETRIC_IDENTIFIER": 5,
78
+ "BLOOD_TYPE": 6,
79
+ "CERTIFICATE_LICENSE_NUMBER": 7,
80
+ "CITY": 8,
81
+ "COMPANY_NAME": 9,
82
+ "COORDINATE": 10,
83
+ "COUNTRY": 11,
84
+ "COUNTY": 12,
85
+ "CREDIT_DEBIT_CARD": 13,
86
+ "CUSTOMER_ID": 14,
87
+ "CVV": 15,
88
+ "DATE": 16,
89
+ "DATE_OF_BIRTH": 17,
90
+ "DATE_TIME": 18,
91
+ "DEVICE_IDENTIFIER": 19,
92
+ "EDUCATION_LEVEL": 20,
93
+ "EMAIL": 21,
94
+ "EMPLOYEE_ID": 22,
95
+ "EMPLOYMENT_STATUS": 23,
96
+ "FAX_NUMBER": 24,
97
+ "FIRST_NAME": 25,
98
+ "GENDER": 26,
99
+ "HEALTH_PLAN_BENEFICIARY_NUMBER": 27,
100
+ "HTTP_COOKIE": 28,
101
+ "IPV4": 29,
102
+ "IPV6": 30,
103
+ "LANGUAGE": 31,
104
+ "LAST_NAME": 32,
105
+ "LICENSE_PLATE": 33,
106
+ "MAC_ADDRESS": 34,
107
+ "MEDICAL_RECORD_NUMBER": 35,
108
+ "NATIONAL_ID": 36,
109
+ "O": 0,
110
+ "OCCUPATION": 37,
111
+ "PASSWORD": 38,
112
+ "PHONE_NUMBER": 39,
113
+ "PIN": 40,
114
+ "POLITICAL_VIEW": 41,
115
+ "POSTCODE": 42,
116
+ "RACE_ETHNICITY": 43,
117
+ "RELIGIOUS_BELIEF": 44,
118
+ "SEXUALITY": 45,
119
+ "SSN": 46,
120
+ "STATE": 47,
121
+ "STREET_ADDRESS": 48,
122
+ "SWIFT_BIC": 49,
123
+ "TAX_ID": 50,
124
+ "TIME": 51,
125
+ "UNIQUE_ID": 52,
126
+ "URL": 53,
127
+ "USER_NAME": 54,
128
+ "VEHICLE_IDENTIFIER": 55
129
+ },
130
+ "layer_norm_eps": 1e-07,
131
+ "legacy": true,
132
+ "max_position_embeddings": 512,
133
+ "max_relative_positions": -1,
134
+ "model_type": "deberta-v2",
135
+ "norm_rel_ebd": "layer_norm",
136
+ "num_attention_heads": 12,
137
+ "num_hidden_layers": 12,
138
+ "pad_token_id": 0,
139
+ "pooler_dropout": 0,
140
+ "pooler_hidden_act": "gelu",
141
+ "pooler_hidden_size": 768,
142
+ "pos_att_type": [
143
+ "p2c",
144
+ "c2p"
145
+ ],
146
+ "position_biased_input": false,
147
+ "position_buckets": 256,
148
+ "relative_attention": true,
149
+ "share_att_key": true,
150
+ "tie_word_embeddings": true,
151
+ "transformers_version": "5.3.0",
152
+ "type_vocab_size": 0,
153
+ "use_cache": false,
154
+ "vocab_size": 128100
155
+ }
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87a0354a0db6fdc28c05d17210b207890120d744297ec9e4f96767d8b9608370
3
+ size 741501240
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd5e5ac5ae12351dad9783541b4be8fa7b4bd344ea52c4314e9f5a71d9f3d4bc
3
+ size 1483126155
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:018221c6048850691506bbeb014c39f7c0a12c29a451580ddb7832ee17d0b4b2
3
+ size 14645
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53aabee4a46815e07329f746a41b34271f592e9b883f5dac0650c99884e10eb1
3
+ size 1465
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/tokenizer_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "backend": "tokenizers",
4
+ "bos_token": "[CLS]",
5
+ "cls_token": "[CLS]",
6
+ "do_lower_case": false,
7
+ "eos_token": "[SEP]",
8
+ "extra_special_tokens": [
9
+ "[PAD]",
10
+ "[CLS]",
11
+ "[SEP]"
12
+ ],
13
+ "is_local": false,
14
+ "mask_token": "[MASK]",
15
+ "model_max_length": 1000000000000000019884624838656,
16
+ "pad_token": "[PAD]",
17
+ "sep_token": "[SEP]",
18
+ "split_by_punct": false,
19
+ "tokenizer_class": "DebertaV2Tokenizer",
20
+ "unk_id": 3,
21
+ "unk_token": "[UNK]",
22
+ "vocab_type": "spm"
23
+ }
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/trainer_state.json ADDED
@@ -0,0 +1,181 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.6800000000000002,
6
+ "eval_steps": 500,
7
+ "global_step": 10500,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.08,
14
+ "grad_norm": 0.003990198019891977,
15
+ "learning_rate": 1.9201600000000003e-05,
16
+ "loss": 0.005083163738250732,
17
+ "step": 500
18
+ },
19
+ {
20
+ "epoch": 0.16,
21
+ "grad_norm": 0.006030593533068895,
22
+ "learning_rate": 1.84016e-05,
23
+ "loss": 0.0017981358766555786,
24
+ "step": 1000
25
+ },
26
+ {
27
+ "epoch": 0.24,
28
+ "grad_norm": 0.0030848984606564045,
29
+ "learning_rate": 1.76016e-05,
30
+ "loss": 0.0008149439096450806,
31
+ "step": 1500
32
+ },
33
+ {
34
+ "epoch": 0.32,
35
+ "grad_norm": 0.002702153753489256,
36
+ "learning_rate": 1.6801600000000002e-05,
37
+ "loss": 0.0004432150423526764,
38
+ "step": 2000
39
+ },
40
+ {
41
+ "epoch": 0.4,
42
+ "grad_norm": 0.002467973157763481,
43
+ "learning_rate": 1.6001600000000003e-05,
44
+ "loss": 0.00028532958030700684,
45
+ "step": 2500
46
+ },
47
+ {
48
+ "epoch": 0.48,
49
+ "grad_norm": 0.0021964123006910086,
50
+ "learning_rate": 1.52016e-05,
51
+ "loss": 0.00020916727185249327,
52
+ "step": 3000
53
+ },
54
+ {
55
+ "epoch": 0.56,
56
+ "grad_norm": 0.004308697767555714,
57
+ "learning_rate": 1.4401600000000003e-05,
58
+ "loss": 0.00017732538282871246,
59
+ "step": 3500
60
+ },
61
+ {
62
+ "epoch": 0.64,
63
+ "grad_norm": 0.002398895798251033,
64
+ "learning_rate": 1.3601600000000002e-05,
65
+ "loss": 0.000154940202832222,
66
+ "step": 4000
67
+ },
68
+ {
69
+ "epoch": 0.72,
70
+ "grad_norm": 0.0022735886741429567,
71
+ "learning_rate": 1.2801600000000001e-05,
72
+ "loss": 0.00014079591631889344,
73
+ "step": 4500
74
+ },
75
+ {
76
+ "epoch": 0.8,
77
+ "grad_norm": 0.00556814344599843,
78
+ "learning_rate": 1.20016e-05,
79
+ "loss": 0.0001373387724161148,
80
+ "step": 5000
81
+ },
82
+ {
83
+ "epoch": 0.88,
84
+ "grad_norm": 0.0012896452099084854,
85
+ "learning_rate": 1.1201600000000001e-05,
86
+ "loss": 0.00012117443978786468,
87
+ "step": 5500
88
+ },
89
+ {
90
+ "epoch": 0.96,
91
+ "grad_norm": 0.0033518935088068247,
92
+ "learning_rate": 1.04016e-05,
93
+ "loss": 0.00011866840720176697,
94
+ "step": 6000
95
+ },
96
+ {
97
+ "epoch": 1.04,
98
+ "grad_norm": 0.0012641055509448051,
99
+ "learning_rate": 9.601600000000001e-06,
100
+ "loss": 0.0001060236245393753,
101
+ "step": 6500
102
+ },
103
+ {
104
+ "epoch": 1.12,
105
+ "grad_norm": 0.000537463347427547,
106
+ "learning_rate": 8.8016e-06,
107
+ "loss": 0.00010027290880680084,
108
+ "step": 7000
109
+ },
110
+ {
111
+ "epoch": 1.2,
112
+ "grad_norm": 0.0018809024477377534,
113
+ "learning_rate": 8.0016e-06,
114
+ "loss": 9.29199755191803e-05,
115
+ "step": 7500
116
+ },
117
+ {
118
+ "epoch": 1.28,
119
+ "grad_norm": 0.005723062437027693,
120
+ "learning_rate": 7.201600000000001e-06,
121
+ "loss": 9.09605547785759e-05,
122
+ "step": 8000
123
+ },
124
+ {
125
+ "epoch": 1.3599999999999999,
126
+ "grad_norm": 0.0005352628650143743,
127
+ "learning_rate": 6.4016e-06,
128
+ "loss": 9.268201887607574e-05,
129
+ "step": 8500
130
+ },
131
+ {
132
+ "epoch": 1.44,
133
+ "grad_norm": 0.001981391105800867,
134
+ "learning_rate": 5.601600000000001e-06,
135
+ "loss": 8.78596231341362e-05,
136
+ "step": 9000
137
+ },
138
+ {
139
+ "epoch": 1.52,
140
+ "grad_norm": 0.007651986088603735,
141
+ "learning_rate": 4.8016e-06,
142
+ "loss": 8.183776587247848e-05,
143
+ "step": 9500
144
+ },
145
+ {
146
+ "epoch": 1.6,
147
+ "grad_norm": 0.002296850783750415,
148
+ "learning_rate": 4.001600000000001e-06,
149
+ "loss": 8.76401886343956e-05,
150
+ "step": 10000
151
+ },
152
+ {
153
+ "epoch": 1.6800000000000002,
154
+ "grad_norm": 0.0005856395000591874,
155
+ "learning_rate": 3.2016000000000004e-06,
156
+ "loss": 7.893749326467514e-05,
157
+ "step": 10500
158
+ }
159
+ ],
160
+ "logging_steps": 500,
161
+ "max_steps": 12500,
162
+ "num_input_tokens_seen": 0,
163
+ "num_train_epochs": 2,
164
+ "save_steps": 500,
165
+ "stateful_callbacks": {
166
+ "TrainerControl": {
167
+ "args": {
168
+ "should_epoch_stop": false,
169
+ "should_evaluate": false,
170
+ "should_log": false,
171
+ "should_save": true,
172
+ "should_training_stop": false
173
+ },
174
+ "attributes": {}
175
+ }
176
+ },
177
+ "total_flos": 4.280919773294515e+16,
178
+ "train_batch_size": 16,
179
+ "trial_name": null,
180
+ "trial_params": null
181
+ }
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-10500/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8038413655afcd9f5ce084a60e4a073f6b35472739c0d53f5df4cbdca825d93e
3
+ size 5201
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/config.json ADDED
@@ -0,0 +1,155 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "DebertaAdvancedSpanClassifier"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": null,
7
+ "dtype": "float32",
8
+ "eos_token_id": null,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "O",
14
+ "1": "ACCOUNT_NUMBER",
15
+ "2": "AGE",
16
+ "3": "API_KEY",
17
+ "4": "BANK_ROUTING_NUMBER",
18
+ "5": "BIOMETRIC_IDENTIFIER",
19
+ "6": "BLOOD_TYPE",
20
+ "7": "CERTIFICATE_LICENSE_NUMBER",
21
+ "8": "CITY",
22
+ "9": "COMPANY_NAME",
23
+ "10": "COORDINATE",
24
+ "11": "COUNTRY",
25
+ "12": "COUNTY",
26
+ "13": "CREDIT_DEBIT_CARD",
27
+ "14": "CUSTOMER_ID",
28
+ "15": "CVV",
29
+ "16": "DATE",
30
+ "17": "DATE_OF_BIRTH",
31
+ "18": "DATE_TIME",
32
+ "19": "DEVICE_IDENTIFIER",
33
+ "20": "EDUCATION_LEVEL",
34
+ "21": "EMAIL",
35
+ "22": "EMPLOYEE_ID",
36
+ "23": "EMPLOYMENT_STATUS",
37
+ "24": "FAX_NUMBER",
38
+ "25": "FIRST_NAME",
39
+ "26": "GENDER",
40
+ "27": "HEALTH_PLAN_BENEFICIARY_NUMBER",
41
+ "28": "HTTP_COOKIE",
42
+ "29": "IPV4",
43
+ "30": "IPV6",
44
+ "31": "LANGUAGE",
45
+ "32": "LAST_NAME",
46
+ "33": "LICENSE_PLATE",
47
+ "34": "MAC_ADDRESS",
48
+ "35": "MEDICAL_RECORD_NUMBER",
49
+ "36": "NATIONAL_ID",
50
+ "37": "OCCUPATION",
51
+ "38": "PASSWORD",
52
+ "39": "PHONE_NUMBER",
53
+ "40": "PIN",
54
+ "41": "POLITICAL_VIEW",
55
+ "42": "POSTCODE",
56
+ "43": "RACE_ETHNICITY",
57
+ "44": "RELIGIOUS_BELIEF",
58
+ "45": "SEXUALITY",
59
+ "46": "SSN",
60
+ "47": "STATE",
61
+ "48": "STREET_ADDRESS",
62
+ "49": "SWIFT_BIC",
63
+ "50": "TAX_ID",
64
+ "51": "TIME",
65
+ "52": "UNIQUE_ID",
66
+ "53": "URL",
67
+ "54": "USER_NAME",
68
+ "55": "VEHICLE_IDENTIFIER"
69
+ },
70
+ "initializer_range": 0.02,
71
+ "intermediate_size": 3072,
72
+ "label2id": {
73
+ "ACCOUNT_NUMBER": 1,
74
+ "AGE": 2,
75
+ "API_KEY": 3,
76
+ "BANK_ROUTING_NUMBER": 4,
77
+ "BIOMETRIC_IDENTIFIER": 5,
78
+ "BLOOD_TYPE": 6,
79
+ "CERTIFICATE_LICENSE_NUMBER": 7,
80
+ "CITY": 8,
81
+ "COMPANY_NAME": 9,
82
+ "COORDINATE": 10,
83
+ "COUNTRY": 11,
84
+ "COUNTY": 12,
85
+ "CREDIT_DEBIT_CARD": 13,
86
+ "CUSTOMER_ID": 14,
87
+ "CVV": 15,
88
+ "DATE": 16,
89
+ "DATE_OF_BIRTH": 17,
90
+ "DATE_TIME": 18,
91
+ "DEVICE_IDENTIFIER": 19,
92
+ "EDUCATION_LEVEL": 20,
93
+ "EMAIL": 21,
94
+ "EMPLOYEE_ID": 22,
95
+ "EMPLOYMENT_STATUS": 23,
96
+ "FAX_NUMBER": 24,
97
+ "FIRST_NAME": 25,
98
+ "GENDER": 26,
99
+ "HEALTH_PLAN_BENEFICIARY_NUMBER": 27,
100
+ "HTTP_COOKIE": 28,
101
+ "IPV4": 29,
102
+ "IPV6": 30,
103
+ "LANGUAGE": 31,
104
+ "LAST_NAME": 32,
105
+ "LICENSE_PLATE": 33,
106
+ "MAC_ADDRESS": 34,
107
+ "MEDICAL_RECORD_NUMBER": 35,
108
+ "NATIONAL_ID": 36,
109
+ "O": 0,
110
+ "OCCUPATION": 37,
111
+ "PASSWORD": 38,
112
+ "PHONE_NUMBER": 39,
113
+ "PIN": 40,
114
+ "POLITICAL_VIEW": 41,
115
+ "POSTCODE": 42,
116
+ "RACE_ETHNICITY": 43,
117
+ "RELIGIOUS_BELIEF": 44,
118
+ "SEXUALITY": 45,
119
+ "SSN": 46,
120
+ "STATE": 47,
121
+ "STREET_ADDRESS": 48,
122
+ "SWIFT_BIC": 49,
123
+ "TAX_ID": 50,
124
+ "TIME": 51,
125
+ "UNIQUE_ID": 52,
126
+ "URL": 53,
127
+ "USER_NAME": 54,
128
+ "VEHICLE_IDENTIFIER": 55
129
+ },
130
+ "layer_norm_eps": 1e-07,
131
+ "legacy": true,
132
+ "max_position_embeddings": 512,
133
+ "max_relative_positions": -1,
134
+ "model_type": "deberta-v2",
135
+ "norm_rel_ebd": "layer_norm",
136
+ "num_attention_heads": 12,
137
+ "num_hidden_layers": 12,
138
+ "pad_token_id": 0,
139
+ "pooler_dropout": 0,
140
+ "pooler_hidden_act": "gelu",
141
+ "pooler_hidden_size": 768,
142
+ "pos_att_type": [
143
+ "p2c",
144
+ "c2p"
145
+ ],
146
+ "position_biased_input": false,
147
+ "position_buckets": 256,
148
+ "relative_attention": true,
149
+ "share_att_key": true,
150
+ "tie_word_embeddings": true,
151
+ "transformers_version": "5.3.0",
152
+ "type_vocab_size": 0,
153
+ "use_cache": false,
154
+ "vocab_size": 128100
155
+ }
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f58adfcd0ad218a7f82c3f4cb924ab9bf05b41b8d4a408c5468cba02a82a40e
3
+ size 741501240
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0aedc119db191fb45970c6c13bc5ba3731e61d7c0acc97d96d36035c0b9017d4
3
+ size 1483126155
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:362666c0b3b985d7d4da93cbee2e0f5877f92156946dc7bb5365023f62d4f2b9
3
+ size 14645
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0781165a305129e833f93432848bb7c0a4a3e609e73baffd2075894b986ee557
3
+ size 1465
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/tokenizer_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "backend": "tokenizers",
4
+ "bos_token": "[CLS]",
5
+ "cls_token": "[CLS]",
6
+ "do_lower_case": false,
7
+ "eos_token": "[SEP]",
8
+ "extra_special_tokens": [
9
+ "[PAD]",
10
+ "[CLS]",
11
+ "[SEP]"
12
+ ],
13
+ "is_local": false,
14
+ "mask_token": "[MASK]",
15
+ "model_max_length": 1000000000000000019884624838656,
16
+ "pad_token": "[PAD]",
17
+ "sep_token": "[SEP]",
18
+ "split_by_punct": false,
19
+ "tokenizer_class": "DebertaV2Tokenizer",
20
+ "unk_id": 3,
21
+ "unk_token": "[UNK]",
22
+ "vocab_type": "spm"
23
+ }
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/trainer_state.json ADDED
@@ -0,0 +1,188 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.76,
6
+ "eval_steps": 500,
7
+ "global_step": 11000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.08,
14
+ "grad_norm": 0.003990198019891977,
15
+ "learning_rate": 1.9201600000000003e-05,
16
+ "loss": 0.005083163738250732,
17
+ "step": 500
18
+ },
19
+ {
20
+ "epoch": 0.16,
21
+ "grad_norm": 0.006030593533068895,
22
+ "learning_rate": 1.84016e-05,
23
+ "loss": 0.0017981358766555786,
24
+ "step": 1000
25
+ },
26
+ {
27
+ "epoch": 0.24,
28
+ "grad_norm": 0.0030848984606564045,
29
+ "learning_rate": 1.76016e-05,
30
+ "loss": 0.0008149439096450806,
31
+ "step": 1500
32
+ },
33
+ {
34
+ "epoch": 0.32,
35
+ "grad_norm": 0.002702153753489256,
36
+ "learning_rate": 1.6801600000000002e-05,
37
+ "loss": 0.0004432150423526764,
38
+ "step": 2000
39
+ },
40
+ {
41
+ "epoch": 0.4,
42
+ "grad_norm": 0.002467973157763481,
43
+ "learning_rate": 1.6001600000000003e-05,
44
+ "loss": 0.00028532958030700684,
45
+ "step": 2500
46
+ },
47
+ {
48
+ "epoch": 0.48,
49
+ "grad_norm": 0.0021964123006910086,
50
+ "learning_rate": 1.52016e-05,
51
+ "loss": 0.00020916727185249327,
52
+ "step": 3000
53
+ },
54
+ {
55
+ "epoch": 0.56,
56
+ "grad_norm": 0.004308697767555714,
57
+ "learning_rate": 1.4401600000000003e-05,
58
+ "loss": 0.00017732538282871246,
59
+ "step": 3500
60
+ },
61
+ {
62
+ "epoch": 0.64,
63
+ "grad_norm": 0.002398895798251033,
64
+ "learning_rate": 1.3601600000000002e-05,
65
+ "loss": 0.000154940202832222,
66
+ "step": 4000
67
+ },
68
+ {
69
+ "epoch": 0.72,
70
+ "grad_norm": 0.0022735886741429567,
71
+ "learning_rate": 1.2801600000000001e-05,
72
+ "loss": 0.00014079591631889344,
73
+ "step": 4500
74
+ },
75
+ {
76
+ "epoch": 0.8,
77
+ "grad_norm": 0.00556814344599843,
78
+ "learning_rate": 1.20016e-05,
79
+ "loss": 0.0001373387724161148,
80
+ "step": 5000
81
+ },
82
+ {
83
+ "epoch": 0.88,
84
+ "grad_norm": 0.0012896452099084854,
85
+ "learning_rate": 1.1201600000000001e-05,
86
+ "loss": 0.00012117443978786468,
87
+ "step": 5500
88
+ },
89
+ {
90
+ "epoch": 0.96,
91
+ "grad_norm": 0.0033518935088068247,
92
+ "learning_rate": 1.04016e-05,
93
+ "loss": 0.00011866840720176697,
94
+ "step": 6000
95
+ },
96
+ {
97
+ "epoch": 1.04,
98
+ "grad_norm": 0.0012641055509448051,
99
+ "learning_rate": 9.601600000000001e-06,
100
+ "loss": 0.0001060236245393753,
101
+ "step": 6500
102
+ },
103
+ {
104
+ "epoch": 1.12,
105
+ "grad_norm": 0.000537463347427547,
106
+ "learning_rate": 8.8016e-06,
107
+ "loss": 0.00010027290880680084,
108
+ "step": 7000
109
+ },
110
+ {
111
+ "epoch": 1.2,
112
+ "grad_norm": 0.0018809024477377534,
113
+ "learning_rate": 8.0016e-06,
114
+ "loss": 9.29199755191803e-05,
115
+ "step": 7500
116
+ },
117
+ {
118
+ "epoch": 1.28,
119
+ "grad_norm": 0.005723062437027693,
120
+ "learning_rate": 7.201600000000001e-06,
121
+ "loss": 9.09605547785759e-05,
122
+ "step": 8000
123
+ },
124
+ {
125
+ "epoch": 1.3599999999999999,
126
+ "grad_norm": 0.0005352628650143743,
127
+ "learning_rate": 6.4016e-06,
128
+ "loss": 9.268201887607574e-05,
129
+ "step": 8500
130
+ },
131
+ {
132
+ "epoch": 1.44,
133
+ "grad_norm": 0.001981391105800867,
134
+ "learning_rate": 5.601600000000001e-06,
135
+ "loss": 8.78596231341362e-05,
136
+ "step": 9000
137
+ },
138
+ {
139
+ "epoch": 1.52,
140
+ "grad_norm": 0.007651986088603735,
141
+ "learning_rate": 4.8016e-06,
142
+ "loss": 8.183776587247848e-05,
143
+ "step": 9500
144
+ },
145
+ {
146
+ "epoch": 1.6,
147
+ "grad_norm": 0.002296850783750415,
148
+ "learning_rate": 4.001600000000001e-06,
149
+ "loss": 8.76401886343956e-05,
150
+ "step": 10000
151
+ },
152
+ {
153
+ "epoch": 1.6800000000000002,
154
+ "grad_norm": 0.0005856395000591874,
155
+ "learning_rate": 3.2016000000000004e-06,
156
+ "loss": 7.893749326467514e-05,
157
+ "step": 10500
158
+ },
159
+ {
160
+ "epoch": 1.76,
161
+ "grad_norm": 0.0027157405856996775,
162
+ "learning_rate": 2.4016000000000004e-06,
163
+ "loss": 7.564106583595275e-05,
164
+ "step": 11000
165
+ }
166
+ ],
167
+ "logging_steps": 500,
168
+ "max_steps": 12500,
169
+ "num_input_tokens_seen": 0,
170
+ "num_train_epochs": 2,
171
+ "save_steps": 500,
172
+ "stateful_callbacks": {
173
+ "TrainerControl": {
174
+ "args": {
175
+ "should_epoch_stop": false,
176
+ "should_evaluate": false,
177
+ "should_log": false,
178
+ "should_save": true,
179
+ "should_training_stop": false
180
+ },
181
+ "attributes": {}
182
+ }
183
+ },
184
+ "total_flos": 4.484927012990746e+16,
185
+ "train_batch_size": 16,
186
+ "trial_name": null,
187
+ "trial_params": null
188
+ }
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8038413655afcd9f5ce084a60e4a073f6b35472739c0d53f5df4cbdca825d93e
3
+ size 5201
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/config.json ADDED
@@ -0,0 +1,155 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "DebertaAdvancedSpanClassifier"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": null,
7
+ "dtype": "float32",
8
+ "eos_token_id": null,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "O",
14
+ "1": "ACCOUNT_NUMBER",
15
+ "2": "AGE",
16
+ "3": "API_KEY",
17
+ "4": "BANK_ROUTING_NUMBER",
18
+ "5": "BIOMETRIC_IDENTIFIER",
19
+ "6": "BLOOD_TYPE",
20
+ "7": "CERTIFICATE_LICENSE_NUMBER",
21
+ "8": "CITY",
22
+ "9": "COMPANY_NAME",
23
+ "10": "COORDINATE",
24
+ "11": "COUNTRY",
25
+ "12": "COUNTY",
26
+ "13": "CREDIT_DEBIT_CARD",
27
+ "14": "CUSTOMER_ID",
28
+ "15": "CVV",
29
+ "16": "DATE",
30
+ "17": "DATE_OF_BIRTH",
31
+ "18": "DATE_TIME",
32
+ "19": "DEVICE_IDENTIFIER",
33
+ "20": "EDUCATION_LEVEL",
34
+ "21": "EMAIL",
35
+ "22": "EMPLOYEE_ID",
36
+ "23": "EMPLOYMENT_STATUS",
37
+ "24": "FAX_NUMBER",
38
+ "25": "FIRST_NAME",
39
+ "26": "GENDER",
40
+ "27": "HEALTH_PLAN_BENEFICIARY_NUMBER",
41
+ "28": "HTTP_COOKIE",
42
+ "29": "IPV4",
43
+ "30": "IPV6",
44
+ "31": "LANGUAGE",
45
+ "32": "LAST_NAME",
46
+ "33": "LICENSE_PLATE",
47
+ "34": "MAC_ADDRESS",
48
+ "35": "MEDICAL_RECORD_NUMBER",
49
+ "36": "NATIONAL_ID",
50
+ "37": "OCCUPATION",
51
+ "38": "PASSWORD",
52
+ "39": "PHONE_NUMBER",
53
+ "40": "PIN",
54
+ "41": "POLITICAL_VIEW",
55
+ "42": "POSTCODE",
56
+ "43": "RACE_ETHNICITY",
57
+ "44": "RELIGIOUS_BELIEF",
58
+ "45": "SEXUALITY",
59
+ "46": "SSN",
60
+ "47": "STATE",
61
+ "48": "STREET_ADDRESS",
62
+ "49": "SWIFT_BIC",
63
+ "50": "TAX_ID",
64
+ "51": "TIME",
65
+ "52": "UNIQUE_ID",
66
+ "53": "URL",
67
+ "54": "USER_NAME",
68
+ "55": "VEHICLE_IDENTIFIER"
69
+ },
70
+ "initializer_range": 0.02,
71
+ "intermediate_size": 3072,
72
+ "label2id": {
73
+ "ACCOUNT_NUMBER": 1,
74
+ "AGE": 2,
75
+ "API_KEY": 3,
76
+ "BANK_ROUTING_NUMBER": 4,
77
+ "BIOMETRIC_IDENTIFIER": 5,
78
+ "BLOOD_TYPE": 6,
79
+ "CERTIFICATE_LICENSE_NUMBER": 7,
80
+ "CITY": 8,
81
+ "COMPANY_NAME": 9,
82
+ "COORDINATE": 10,
83
+ "COUNTRY": 11,
84
+ "COUNTY": 12,
85
+ "CREDIT_DEBIT_CARD": 13,
86
+ "CUSTOMER_ID": 14,
87
+ "CVV": 15,
88
+ "DATE": 16,
89
+ "DATE_OF_BIRTH": 17,
90
+ "DATE_TIME": 18,
91
+ "DEVICE_IDENTIFIER": 19,
92
+ "EDUCATION_LEVEL": 20,
93
+ "EMAIL": 21,
94
+ "EMPLOYEE_ID": 22,
95
+ "EMPLOYMENT_STATUS": 23,
96
+ "FAX_NUMBER": 24,
97
+ "FIRST_NAME": 25,
98
+ "GENDER": 26,
99
+ "HEALTH_PLAN_BENEFICIARY_NUMBER": 27,
100
+ "HTTP_COOKIE": 28,
101
+ "IPV4": 29,
102
+ "IPV6": 30,
103
+ "LANGUAGE": 31,
104
+ "LAST_NAME": 32,
105
+ "LICENSE_PLATE": 33,
106
+ "MAC_ADDRESS": 34,
107
+ "MEDICAL_RECORD_NUMBER": 35,
108
+ "NATIONAL_ID": 36,
109
+ "O": 0,
110
+ "OCCUPATION": 37,
111
+ "PASSWORD": 38,
112
+ "PHONE_NUMBER": 39,
113
+ "PIN": 40,
114
+ "POLITICAL_VIEW": 41,
115
+ "POSTCODE": 42,
116
+ "RACE_ETHNICITY": 43,
117
+ "RELIGIOUS_BELIEF": 44,
118
+ "SEXUALITY": 45,
119
+ "SSN": 46,
120
+ "STATE": 47,
121
+ "STREET_ADDRESS": 48,
122
+ "SWIFT_BIC": 49,
123
+ "TAX_ID": 50,
124
+ "TIME": 51,
125
+ "UNIQUE_ID": 52,
126
+ "URL": 53,
127
+ "USER_NAME": 54,
128
+ "VEHICLE_IDENTIFIER": 55
129
+ },
130
+ "layer_norm_eps": 1e-07,
131
+ "legacy": true,
132
+ "max_position_embeddings": 512,
133
+ "max_relative_positions": -1,
134
+ "model_type": "deberta-v2",
135
+ "norm_rel_ebd": "layer_norm",
136
+ "num_attention_heads": 12,
137
+ "num_hidden_layers": 12,
138
+ "pad_token_id": 0,
139
+ "pooler_dropout": 0,
140
+ "pooler_hidden_act": "gelu",
141
+ "pooler_hidden_size": 768,
142
+ "pos_att_type": [
143
+ "p2c",
144
+ "c2p"
145
+ ],
146
+ "position_biased_input": false,
147
+ "position_buckets": 256,
148
+ "relative_attention": true,
149
+ "share_att_key": true,
150
+ "tie_word_embeddings": true,
151
+ "transformers_version": "5.3.0",
152
+ "type_vocab_size": 0,
153
+ "use_cache": false,
154
+ "vocab_size": 128100
155
+ }
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72474ed547edfe0e7f61c54ee6e744142b3bff51844b4a0b80a8c01100806848
3
+ size 741501240
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f085c7c84cc8dfa14e0241fad5aae856e807f1fe30077d7b5503fd9de6b4dd2
3
+ size 1483126155
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd33cfdafad4159d073c128dcbba8a9aa03ba2db66d0ea9b5d9a2d0f11fd2575
3
+ size 14645
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c752e6b034370e6ff4da907200224d1bb40961b368f416b4385798c8e2f0c79
3
+ size 1465
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/tokenizer_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "backend": "tokenizers",
4
+ "bos_token": "[CLS]",
5
+ "cls_token": "[CLS]",
6
+ "do_lower_case": false,
7
+ "eos_token": "[SEP]",
8
+ "extra_special_tokens": [
9
+ "[PAD]",
10
+ "[CLS]",
11
+ "[SEP]"
12
+ ],
13
+ "is_local": false,
14
+ "mask_token": "[MASK]",
15
+ "model_max_length": 1000000000000000019884624838656,
16
+ "pad_token": "[PAD]",
17
+ "sep_token": "[SEP]",
18
+ "split_by_punct": false,
19
+ "tokenizer_class": "DebertaV2Tokenizer",
20
+ "unk_id": 3,
21
+ "unk_token": "[UNK]",
22
+ "vocab_type": "spm"
23
+ }
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/trainer_state.json ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.8399999999999999,
6
+ "eval_steps": 500,
7
+ "global_step": 11500,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.08,
14
+ "grad_norm": 0.003990198019891977,
15
+ "learning_rate": 1.9201600000000003e-05,
16
+ "loss": 0.005083163738250732,
17
+ "step": 500
18
+ },
19
+ {
20
+ "epoch": 0.16,
21
+ "grad_norm": 0.006030593533068895,
22
+ "learning_rate": 1.84016e-05,
23
+ "loss": 0.0017981358766555786,
24
+ "step": 1000
25
+ },
26
+ {
27
+ "epoch": 0.24,
28
+ "grad_norm": 0.0030848984606564045,
29
+ "learning_rate": 1.76016e-05,
30
+ "loss": 0.0008149439096450806,
31
+ "step": 1500
32
+ },
33
+ {
34
+ "epoch": 0.32,
35
+ "grad_norm": 0.002702153753489256,
36
+ "learning_rate": 1.6801600000000002e-05,
37
+ "loss": 0.0004432150423526764,
38
+ "step": 2000
39
+ },
40
+ {
41
+ "epoch": 0.4,
42
+ "grad_norm": 0.002467973157763481,
43
+ "learning_rate": 1.6001600000000003e-05,
44
+ "loss": 0.00028532958030700684,
45
+ "step": 2500
46
+ },
47
+ {
48
+ "epoch": 0.48,
49
+ "grad_norm": 0.0021964123006910086,
50
+ "learning_rate": 1.52016e-05,
51
+ "loss": 0.00020916727185249327,
52
+ "step": 3000
53
+ },
54
+ {
55
+ "epoch": 0.56,
56
+ "grad_norm": 0.004308697767555714,
57
+ "learning_rate": 1.4401600000000003e-05,
58
+ "loss": 0.00017732538282871246,
59
+ "step": 3500
60
+ },
61
+ {
62
+ "epoch": 0.64,
63
+ "grad_norm": 0.002398895798251033,
64
+ "learning_rate": 1.3601600000000002e-05,
65
+ "loss": 0.000154940202832222,
66
+ "step": 4000
67
+ },
68
+ {
69
+ "epoch": 0.72,
70
+ "grad_norm": 0.0022735886741429567,
71
+ "learning_rate": 1.2801600000000001e-05,
72
+ "loss": 0.00014079591631889344,
73
+ "step": 4500
74
+ },
75
+ {
76
+ "epoch": 0.8,
77
+ "grad_norm": 0.00556814344599843,
78
+ "learning_rate": 1.20016e-05,
79
+ "loss": 0.0001373387724161148,
80
+ "step": 5000
81
+ },
82
+ {
83
+ "epoch": 0.88,
84
+ "grad_norm": 0.0012896452099084854,
85
+ "learning_rate": 1.1201600000000001e-05,
86
+ "loss": 0.00012117443978786468,
87
+ "step": 5500
88
+ },
89
+ {
90
+ "epoch": 0.96,
91
+ "grad_norm": 0.0033518935088068247,
92
+ "learning_rate": 1.04016e-05,
93
+ "loss": 0.00011866840720176697,
94
+ "step": 6000
95
+ },
96
+ {
97
+ "epoch": 1.04,
98
+ "grad_norm": 0.0012641055509448051,
99
+ "learning_rate": 9.601600000000001e-06,
100
+ "loss": 0.0001060236245393753,
101
+ "step": 6500
102
+ },
103
+ {
104
+ "epoch": 1.12,
105
+ "grad_norm": 0.000537463347427547,
106
+ "learning_rate": 8.8016e-06,
107
+ "loss": 0.00010027290880680084,
108
+ "step": 7000
109
+ },
110
+ {
111
+ "epoch": 1.2,
112
+ "grad_norm": 0.0018809024477377534,
113
+ "learning_rate": 8.0016e-06,
114
+ "loss": 9.29199755191803e-05,
115
+ "step": 7500
116
+ },
117
+ {
118
+ "epoch": 1.28,
119
+ "grad_norm": 0.005723062437027693,
120
+ "learning_rate": 7.201600000000001e-06,
121
+ "loss": 9.09605547785759e-05,
122
+ "step": 8000
123
+ },
124
+ {
125
+ "epoch": 1.3599999999999999,
126
+ "grad_norm": 0.0005352628650143743,
127
+ "learning_rate": 6.4016e-06,
128
+ "loss": 9.268201887607574e-05,
129
+ "step": 8500
130
+ },
131
+ {
132
+ "epoch": 1.44,
133
+ "grad_norm": 0.001981391105800867,
134
+ "learning_rate": 5.601600000000001e-06,
135
+ "loss": 8.78596231341362e-05,
136
+ "step": 9000
137
+ },
138
+ {
139
+ "epoch": 1.52,
140
+ "grad_norm": 0.007651986088603735,
141
+ "learning_rate": 4.8016e-06,
142
+ "loss": 8.183776587247848e-05,
143
+ "step": 9500
144
+ },
145
+ {
146
+ "epoch": 1.6,
147
+ "grad_norm": 0.002296850783750415,
148
+ "learning_rate": 4.001600000000001e-06,
149
+ "loss": 8.76401886343956e-05,
150
+ "step": 10000
151
+ },
152
+ {
153
+ "epoch": 1.6800000000000002,
154
+ "grad_norm": 0.0005856395000591874,
155
+ "learning_rate": 3.2016000000000004e-06,
156
+ "loss": 7.893749326467514e-05,
157
+ "step": 10500
158
+ },
159
+ {
160
+ "epoch": 1.76,
161
+ "grad_norm": 0.0027157405856996775,
162
+ "learning_rate": 2.4016000000000004e-06,
163
+ "loss": 7.564106583595275e-05,
164
+ "step": 11000
165
+ },
166
+ {
167
+ "epoch": 1.8399999999999999,
168
+ "grad_norm": 0.0008732234709896147,
169
+ "learning_rate": 1.6016e-06,
170
+ "loss": 7.321659475564957e-05,
171
+ "step": 11500
172
+ }
173
+ ],
174
+ "logging_steps": 500,
175
+ "max_steps": 12500,
176
+ "num_input_tokens_seen": 0,
177
+ "num_train_epochs": 2,
178
+ "save_steps": 500,
179
+ "stateful_callbacks": {
180
+ "TrainerControl": {
181
+ "args": {
182
+ "should_epoch_stop": false,
183
+ "should_evaluate": false,
184
+ "should_log": false,
185
+ "should_save": true,
186
+ "should_training_stop": false
187
+ },
188
+ "attributes": {}
189
+ }
190
+ },
191
+ "total_flos": 4.689642374517274e+16,
192
+ "train_batch_size": 16,
193
+ "trial_name": null,
194
+ "trial_params": null
195
+ }
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-11500/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8038413655afcd9f5ce084a60e4a073f6b35472739c0d53f5df4cbdca825d93e
3
+ size 5201
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-12000/config.json ADDED
@@ -0,0 +1,155 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "DebertaAdvancedSpanClassifier"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": null,
7
+ "dtype": "float32",
8
+ "eos_token_id": null,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "O",
14
+ "1": "ACCOUNT_NUMBER",
15
+ "2": "AGE",
16
+ "3": "API_KEY",
17
+ "4": "BANK_ROUTING_NUMBER",
18
+ "5": "BIOMETRIC_IDENTIFIER",
19
+ "6": "BLOOD_TYPE",
20
+ "7": "CERTIFICATE_LICENSE_NUMBER",
21
+ "8": "CITY",
22
+ "9": "COMPANY_NAME",
23
+ "10": "COORDINATE",
24
+ "11": "COUNTRY",
25
+ "12": "COUNTY",
26
+ "13": "CREDIT_DEBIT_CARD",
27
+ "14": "CUSTOMER_ID",
28
+ "15": "CVV",
29
+ "16": "DATE",
30
+ "17": "DATE_OF_BIRTH",
31
+ "18": "DATE_TIME",
32
+ "19": "DEVICE_IDENTIFIER",
33
+ "20": "EDUCATION_LEVEL",
34
+ "21": "EMAIL",
35
+ "22": "EMPLOYEE_ID",
36
+ "23": "EMPLOYMENT_STATUS",
37
+ "24": "FAX_NUMBER",
38
+ "25": "FIRST_NAME",
39
+ "26": "GENDER",
40
+ "27": "HEALTH_PLAN_BENEFICIARY_NUMBER",
41
+ "28": "HTTP_COOKIE",
42
+ "29": "IPV4",
43
+ "30": "IPV6",
44
+ "31": "LANGUAGE",
45
+ "32": "LAST_NAME",
46
+ "33": "LICENSE_PLATE",
47
+ "34": "MAC_ADDRESS",
48
+ "35": "MEDICAL_RECORD_NUMBER",
49
+ "36": "NATIONAL_ID",
50
+ "37": "OCCUPATION",
51
+ "38": "PASSWORD",
52
+ "39": "PHONE_NUMBER",
53
+ "40": "PIN",
54
+ "41": "POLITICAL_VIEW",
55
+ "42": "POSTCODE",
56
+ "43": "RACE_ETHNICITY",
57
+ "44": "RELIGIOUS_BELIEF",
58
+ "45": "SEXUALITY",
59
+ "46": "SSN",
60
+ "47": "STATE",
61
+ "48": "STREET_ADDRESS",
62
+ "49": "SWIFT_BIC",
63
+ "50": "TAX_ID",
64
+ "51": "TIME",
65
+ "52": "UNIQUE_ID",
66
+ "53": "URL",
67
+ "54": "USER_NAME",
68
+ "55": "VEHICLE_IDENTIFIER"
69
+ },
70
+ "initializer_range": 0.02,
71
+ "intermediate_size": 3072,
72
+ "label2id": {
73
+ "ACCOUNT_NUMBER": 1,
74
+ "AGE": 2,
75
+ "API_KEY": 3,
76
+ "BANK_ROUTING_NUMBER": 4,
77
+ "BIOMETRIC_IDENTIFIER": 5,
78
+ "BLOOD_TYPE": 6,
79
+ "CERTIFICATE_LICENSE_NUMBER": 7,
80
+ "CITY": 8,
81
+ "COMPANY_NAME": 9,
82
+ "COORDINATE": 10,
83
+ "COUNTRY": 11,
84
+ "COUNTY": 12,
85
+ "CREDIT_DEBIT_CARD": 13,
86
+ "CUSTOMER_ID": 14,
87
+ "CVV": 15,
88
+ "DATE": 16,
89
+ "DATE_OF_BIRTH": 17,
90
+ "DATE_TIME": 18,
91
+ "DEVICE_IDENTIFIER": 19,
92
+ "EDUCATION_LEVEL": 20,
93
+ "EMAIL": 21,
94
+ "EMPLOYEE_ID": 22,
95
+ "EMPLOYMENT_STATUS": 23,
96
+ "FAX_NUMBER": 24,
97
+ "FIRST_NAME": 25,
98
+ "GENDER": 26,
99
+ "HEALTH_PLAN_BENEFICIARY_NUMBER": 27,
100
+ "HTTP_COOKIE": 28,
101
+ "IPV4": 29,
102
+ "IPV6": 30,
103
+ "LANGUAGE": 31,
104
+ "LAST_NAME": 32,
105
+ "LICENSE_PLATE": 33,
106
+ "MAC_ADDRESS": 34,
107
+ "MEDICAL_RECORD_NUMBER": 35,
108
+ "NATIONAL_ID": 36,
109
+ "O": 0,
110
+ "OCCUPATION": 37,
111
+ "PASSWORD": 38,
112
+ "PHONE_NUMBER": 39,
113
+ "PIN": 40,
114
+ "POLITICAL_VIEW": 41,
115
+ "POSTCODE": 42,
116
+ "RACE_ETHNICITY": 43,
117
+ "RELIGIOUS_BELIEF": 44,
118
+ "SEXUALITY": 45,
119
+ "SSN": 46,
120
+ "STATE": 47,
121
+ "STREET_ADDRESS": 48,
122
+ "SWIFT_BIC": 49,
123
+ "TAX_ID": 50,
124
+ "TIME": 51,
125
+ "UNIQUE_ID": 52,
126
+ "URL": 53,
127
+ "USER_NAME": 54,
128
+ "VEHICLE_IDENTIFIER": 55
129
+ },
130
+ "layer_norm_eps": 1e-07,
131
+ "legacy": true,
132
+ "max_position_embeddings": 512,
133
+ "max_relative_positions": -1,
134
+ "model_type": "deberta-v2",
135
+ "norm_rel_ebd": "layer_norm",
136
+ "num_attention_heads": 12,
137
+ "num_hidden_layers": 12,
138
+ "pad_token_id": 0,
139
+ "pooler_dropout": 0,
140
+ "pooler_hidden_act": "gelu",
141
+ "pooler_hidden_size": 768,
142
+ "pos_att_type": [
143
+ "p2c",
144
+ "c2p"
145
+ ],
146
+ "position_biased_input": false,
147
+ "position_buckets": 256,
148
+ "relative_attention": true,
149
+ "share_att_key": true,
150
+ "tie_word_embeddings": true,
151
+ "transformers_version": "5.3.0",
152
+ "type_vocab_size": 0,
153
+ "use_cache": false,
154
+ "vocab_size": 128100
155
+ }
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-12000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5af02cee80769f08525678aca8d8c38e2b6f4016983eb805e9eb67c1eb32f738
3
+ size 741501240
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-12000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:524a03d7ac37d51c5ccf73c947cf330ddafc0d28ba2afe0403763be20c950447
3
+ size 1483126155
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-12000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40c21cda331e953a74503f2b5752c0cc000af50420552d8db841381340ae9c5d
3
+ size 14645
nemotron-pii-ready/SpanBased-CustomDeBERTa/checkpoint-12000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5cc3711b605e3facd6e23f2dcd372873fd5e33f45d714cf3927ae7570548fc3
3
+ size 1465