jjstuart commited on
Commit
861e0e8
·
verified ·
1 Parent(s): 6b9da8a

Upload folder using huggingface_hub

Browse files
Files changed (38) hide show
  1. README.md +62 -67
  2. checkpoints/checkpoint-1328/1_Pooling/config.json +10 -0
  3. checkpoints/checkpoint-1328/README.md +466 -0
  4. checkpoints/checkpoint-1328/config.json +25 -0
  5. checkpoints/checkpoint-1328/config_sentence_transformers.json +14 -0
  6. checkpoints/checkpoint-1328/model.safetensors +3 -0
  7. checkpoints/checkpoint-1328/modules.json +20 -0
  8. checkpoints/checkpoint-1328/optimizer.pt +3 -0
  9. checkpoints/checkpoint-1328/rng_state.pth +3 -0
  10. checkpoints/checkpoint-1328/scheduler.pt +3 -0
  11. checkpoints/checkpoint-1328/sentence_bert_config.json +4 -0
  12. checkpoints/checkpoint-1328/special_tokens_map.json +37 -0
  13. checkpoints/checkpoint-1328/tokenizer.json +0 -0
  14. checkpoints/checkpoint-1328/tokenizer_config.json +65 -0
  15. checkpoints/checkpoint-1328/trainer_state.json +112 -0
  16. checkpoints/checkpoint-1328/training_args.bin +3 -0
  17. checkpoints/checkpoint-1328/vocab.txt +0 -0
  18. checkpoints/checkpoint-1660/1_Pooling/config.json +10 -0
  19. checkpoints/checkpoint-1660/README.md +469 -0
  20. checkpoints/checkpoint-1660/config.json +25 -0
  21. checkpoints/checkpoint-1660/config_sentence_transformers.json +14 -0
  22. checkpoints/checkpoint-1660/model.safetensors +3 -0
  23. checkpoints/checkpoint-1660/modules.json +20 -0
  24. checkpoints/checkpoint-1660/optimizer.pt +3 -0
  25. checkpoints/checkpoint-1660/rng_state.pth +3 -0
  26. checkpoints/checkpoint-1660/scheduler.pt +3 -0
  27. checkpoints/checkpoint-1660/sentence_bert_config.json +4 -0
  28. checkpoints/checkpoint-1660/special_tokens_map.json +37 -0
  29. checkpoints/checkpoint-1660/tokenizer.json +0 -0
  30. checkpoints/checkpoint-1660/tokenizer_config.json +65 -0
  31. checkpoints/checkpoint-1660/trainer_state.json +135 -0
  32. checkpoints/checkpoint-1660/training_args.bin +3 -0
  33. checkpoints/checkpoint-1660/vocab.txt +0 -0
  34. checkpoints/eval/triplet_evaluation_retrieval-eval_results.csv +10 -25
  35. checkpoints/runs/Dec13_16-18-12_rego-trainer-0/events.out.tfevents.1765642693.rego-trainer-0.4819.0 +3 -0
  36. eval/triplet_evaluation_retrieval-eval_results.csv +5 -12
  37. model.safetensors +1 -1
  38. training_info.json +2 -2
README.md CHANGED
@@ -5,20 +5,21 @@ tags:
5
  - feature-extraction
6
  - dense
7
  - generated_from_trainer
8
- - dataset_size:36065
9
  - loss:TripletLoss
10
  base_model: sentence-transformers/all-MiniLM-L6-v2
11
  widget:
12
- - source_sentence: Check that sbom component has empty version field.
 
13
  sentences:
14
- - 'Helper: lib.k8s.name_version
15
 
16
- Signature: name_version(resource)
17
 
18
  Description: '
19
- - 'Helper: lib.k8s.name
20
 
21
- Signature: name(resource)
22
 
23
  Description: '
24
  - 'Helper: lib.k8s.name
@@ -26,84 +27,78 @@ widget:
26
  Signature: name(resource)
27
 
28
  Description: '
29
- - source_sentence: How can I verify that an image manifest is accessible before allowing
30
- the operation to proceed?
31
  sentences:
32
- - 'Helper: lib.result_helper_with_term
33
 
34
- Signature: result_helper_with_term(chain, failure_sprintf_params, term)
35
 
36
  Description: '
37
- - 'Helper: lib.to_set
38
 
39
- Signature: to_set(arr)
40
 
41
  Description: '
42
- - 'Helper: lib.result_helper
43
 
44
- Signature: result_helper(chain, failure_sprintf_params)
45
 
46
  Description: '
47
- - source_sentence: verify that how can i verify that an attestation created by the
48
- rhtap multi-ci build pipeline is present?
 
49
  sentences:
50
- - 'Helper: lib.k8s.version
51
 
52
- Signature: version(resource)
53
 
54
  Description: '
55
- - 'Helper: lib.k8s.name
56
 
57
- Signature: name(resource)
58
 
59
  Description: '
60
- - 'Helper: lib.tekton.tasks
61
 
62
- Signature: tasks(obj)
63
 
64
  Description: '
65
- - source_sentence: policy for create a rego deny rule to the tekton task used specifies
66
- an invalid pipeline. the task is annotated with `build.appstudio.redhat.com/pipeline`
67
- annotation, which must be in the set of `allowed_rpm_build_pipelines` in the rule
68
- data
69
  sentences:
70
- - 'Helper: lib.tekton.tasks
71
 
72
- Signature: tasks(obj)
73
 
74
  Description: '
75
- - 'Helper: lib.tekton.untagged_task_references
76
 
77
- Signature: untagged_task_references(tasks)
78
 
79
  Description: '
80
- - 'Helper: lib.pipelinerun_attestations
81
 
82
- Signature: pipelinerun_attestations
83
 
84
  Description: '
85
- - source_sentence: ensure create a rule that verifies all tekton tasks use the latest
86
- known task reference and reports warnings based on the task_expiry_warning_days
87
- configuration setting.
88
  sentences:
89
- - 'Helper: lib.tekton.bundle
90
 
91
- Signature: bundle(task)
92
 
93
  Description: '
94
- - 'Helper: lib.tekton.untagged_task_references
95
 
96
- Signature: untagged_task_references(tasks)
97
 
98
  Description: '
99
- - 'Path: $.subject[*].digest.sha256
100
-
101
- Description: SHA256 digest of the built artifact (hex-encoded, 64 chars). Used
102
- to verify artifact integrity
103
 
104
- Keywords: sha256, digest, hash, artifact integrity, verification, image digest
105
 
106
- Attestation: slsa_provenance_v02'
107
  pipeline_tag: sentence-similarity
108
  library_name: sentence-transformers
109
  metrics:
@@ -119,7 +114,7 @@ model-index:
119
  type: retrieval-eval
120
  metrics:
121
  - type: cosine_accuracy
122
- value: 0.9762973785400391
123
  name: Cosine Accuracy
124
  ---
125
 
@@ -173,9 +168,9 @@ from sentence_transformers import SentenceTransformer
173
  model = SentenceTransformer("sentence_transformers_model_id")
174
  # Run inference
175
  sentences = [
176
- 'ensure create a rule that verifies all tekton tasks use the latest known task reference and reports warnings based on the task_expiry_warning_days configuration setting.',
177
- 'Helper: lib.tekton.bundle\nSignature: bundle(task)\nDescription: ',
178
- 'Helper: lib.tekton.untagged_task_references\nSignature: untagged_task_references(tasks)\nDescription: ',
179
  ]
180
  embeddings = model.encode(sentences)
181
  print(embeddings.shape)
@@ -184,9 +179,9 @@ print(embeddings.shape)
184
  # Get the similarity scores for the embeddings
185
  similarities = model.similarity(embeddings, embeddings)
186
  print(similarities)
187
- # tensor([[ 1.0000, 0.2491, -0.6356],
188
- # [ 0.2491, 1.0000, -0.2850],
189
- # [-0.6356, -0.2850, 1.0000]])
190
  ```
191
 
192
  <!--
@@ -224,7 +219,7 @@ You can finetune this model on your own dataset.
224
 
225
  | Metric | Value |
226
  |:--------------------|:-----------|
227
- | **cosine_accuracy** | **0.9763** |
228
 
229
  <!--
230
  ## Bias, Risks and Limitations
@@ -244,19 +239,19 @@ You can finetune this model on your own dataset.
244
 
245
  #### Unnamed Dataset
246
 
247
- * Size: 36,065 training samples
248
  * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
249
  * Approximate statistics based on the first 1000 samples:
250
  | | sentence_0 | sentence_1 | sentence_2 |
251
  |:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
252
  | type | string | string | string |
253
- | details | <ul><li>min: 4 tokens</li><li>mean: 31.19 tokens</li><li>max: 161 tokens</li></ul> | <ul><li>min: 21 tokens</li><li>mean: 29.05 tokens</li><li>max: 125 tokens</li></ul> | <ul><li>min: 21 tokens</li><li>mean: 27.65 tokens</li><li>max: 125 tokens</li></ul> |
254
  * Samples:
255
- | sentence_0 | sentence_1 | sentence_2 |
256
- |:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
257
- | <code>I need to for each of the packages fetched by hermeto which define externalreferences, verify they are allowed based on the allowed_package_sources rule data key. by default, allowed_package_sources is empty, which means no components with such references are allowed</code> | <code>Helper: lib.sbom.spdx_sboms<br>Signature: spdx_sboms<br>Description: </code> | <code>Helper: lib.sbom.cyclonedx_sboms<br>Signature: cyclonedx_sboms<br>Description: </code> |
258
- | <code>policy for how can i verify if optional labels are present in an image using the optional_labels or fbc_optional_labels rule data keys?</code> | <code>Helper: lib.image.parse<br>Signature: parse(ref)<br>Description: </code> | <code>Helper: lib.image.str<br>Signature: str(d)<br>Description: </code> |
259
- | <code>create a policy that cyclonedx component is missing bom-ref.</code> | <code>Path: $.components[*].licenses[*].license.id<br>Description: SPDX license ID for the component<br>Keywords: license, sbom, cyclonedx, licensing, compliance, spdx<br>Attestation: cyclonedx_sbom</code> | <code>Path: $.components[*].purl<br>Description: Package URL (purl) for the component. Unique identifier in purl format<br>Keywords: purl, package url, sbom, cyclonedx, identifier<br>Attestation: cyclonedx_sbom</code> |
260
  * Loss: [<code>TripletLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#tripletloss) with these parameters:
261
  ```json
262
  {
@@ -402,13 +397,13 @@ You can finetune this model on your own dataset.
402
  ### Training Logs
403
  | Epoch | Step | Training Loss | retrieval-eval_cosine_accuracy |
404
  |:------:|:----:|:-------------:|:------------------------------:|
405
- | 0.5 | 141 | - | 0.9608 |
406
- | 1.0 | 282 | - | 0.9713 |
407
- | 1.5 | 423 | - | 0.9753 |
408
- | 1.7730 | 500 | 0.0758 | - |
409
- | 2.0 | 564 | - | 0.9750 |
410
- | 2.5 | 705 | - | 0.9748 |
411
- | 3.0 | 846 | - | 0.9763 |
412
 
413
 
414
  ### Framework Versions
 
5
  - feature-extraction
6
  - dense
7
  - generated_from_trainer
8
+ - dataset_size:42459
9
  - loss:TripletLoss
10
  base_model: sentence-transformers/all-MiniLM-L6-v2
11
  widget:
12
+ - source_sentence: policy for how can i verify if a tekton task version is still supported
13
+ by checking for the build.appstudio.redhat.com/expires-on annotation?
14
  sentences:
15
+ - 'Helper: lib.to_array
16
 
17
+ Signature: to_array(s)
18
 
19
  Description: '
20
+ - 'Helper: lib.pipelinerun_attestations
21
 
22
+ Signature: pipelinerun_attestations
23
 
24
  Description: '
25
  - 'Helper: lib.k8s.name
 
27
  Signature: name(resource)
28
 
29
  Description: '
30
+ - source_sentence: how to check attestation is missing statement field.
 
31
  sentences:
32
+ - 'Helper: lib.k8s.name
33
 
34
+ Signature: name(resource)
35
 
36
  Description: '
37
+ - 'Helper: lib.tekton.untrusted_task_refs
38
 
39
+ Signature: untrusted_task_refs(tasks)
40
 
41
  Description: '
42
+ - 'Helper: lib.k8s.version
43
 
44
+ Signature: version(resource)
45
 
46
  Description: '
47
+ - source_sentence: I need to ensure the operators.openshift.io/valid-subscription
48
+ annotation in the ClusterServiceVersion manifest contains a valid JSON encoded
49
+ non-empty array of strings.
50
  sentences:
51
+ - 'Helper: lib.to_array
52
 
53
+ Signature: to_array(s)
54
 
55
  Description: '
56
+ - 'Helper: lib.image.equal_ref
57
 
58
+ Signature: equal_ref(ref1, ref2)
59
 
60
  Description: '
61
+ - 'Helper: lib.result_helper
62
 
63
+ Signature: result_helper(chain, failure_sprintf_params)
64
 
65
  Description: '
66
+ - source_sentence: write a rule to deny approval for an container image with non-unique
67
+ RPM names
 
 
68
  sentences:
69
+ - 'Helper: lib.result_helper
70
 
71
+ Signature: result_helper(chain, failure_sprintf_params)
72
 
73
  Description: '
74
+ - 'Helper: lib.to_set
75
 
76
+ Signature: to_set(arr)
77
 
78
  Description: '
79
+ - 'Helper: lib.rule_data_defaults
80
 
81
+ Signature: rule_data_defaults
82
 
83
  Description: '
84
+ - source_sentence: check if i need to validate that spdx package is an operating system
85
+ component.
 
86
  sentences:
87
+ - 'Helper: lib.to_set
88
 
89
+ Signature: to_set(arr)
90
 
91
  Description: '
92
+ - 'Helper: lib.rule_data_defaults
93
 
94
+ Signature: rule_data_defaults
95
 
96
  Description: '
97
+ - 'Helper: lib.result_helper
 
 
 
98
 
99
+ Signature: result_helper(chain, failure_sprintf_params)
100
 
101
+ Description: '
102
  pipeline_tag: sentence-similarity
103
  library_name: sentence-transformers
104
  metrics:
 
114
  type: retrieval-eval
115
  metrics:
116
  - type: cosine_accuracy
117
+ value: 0.9834675788879395
118
  name: Cosine Accuracy
119
  ---
120
 
 
168
  model = SentenceTransformer("sentence_transformers_model_id")
169
  # Run inference
170
  sentences = [
171
+ 'check if i need to validate that spdx package is an operating system component.',
172
+ 'Helper: lib.result_helper\nSignature: result_helper(chain, failure_sprintf_params)\nDescription: ',
173
+ 'Helper: lib.to_set\nSignature: to_set(arr)\nDescription: ',
174
  ]
175
  embeddings = model.encode(sentences)
176
  print(embeddings.shape)
 
179
  # Get the similarity scores for the embeddings
180
  similarities = model.similarity(embeddings, embeddings)
181
  print(similarities)
182
+ # tensor([[ 1.0000, 0.4979, -0.4443],
183
+ # [ 0.4979, 1.0000, -0.4918],
184
+ # [-0.4443, -0.4918, 1.0000]])
185
  ```
186
 
187
  <!--
 
219
 
220
  | Metric | Value |
221
  |:--------------------|:-----------|
222
+ | **cosine_accuracy** | **0.9835** |
223
 
224
  <!--
225
  ## Bias, Risks and Limitations
 
239
 
240
  #### Unnamed Dataset
241
 
242
+ * Size: 42,459 training samples
243
  * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
244
  * Approximate statistics based on the first 1000 samples:
245
  | | sentence_0 | sentence_1 | sentence_2 |
246
  |:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
247
  | type | string | string | string |
248
+ | details | <ul><li>min: 4 tokens</li><li>mean: 30.48 tokens</li><li>max: 159 tokens</li></ul> | <ul><li>min: 21 tokens</li><li>mean: 29.64 tokens</li><li>max: 125 tokens</li></ul> | <ul><li>min: 21 tokens</li><li>mean: 27.15 tokens</li><li>max: 125 tokens</li></ul> |
249
  * Samples:
250
+ | sentence_0 | sentence_1 | sentence_2 |
251
+ |:-----------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------|
252
+ | <code>I need to ensure that only images from specific registries are used in our policy</code> | <code>Helper: lib.image.str<br>Signature: str(d)<br>Description: </code> | <code>Helper: lib.konflux.is_validating_image_index<br>Signature: is_validating_image_index<br>Description: </code> |
253
+ | <code>check if check warn</code> | <code>Helper: lib.tekton.expiry_of<br>Signature: expiry_of(task)<br>Description: </code> | <code>Helper: lib.tekton.untagged_task_references<br>Signature: untagged_task_references(tasks)<br>Description: </code> |
254
+ | <code>verify that task has an expiry date set.</code> | <code>Helper: lib.tekton.task_param<br>Signature: task_param(task, name)<br>Description: </code> | <code>Helper: lib.tekton.untagged_task_references<br>Signature: untagged_task_references(tasks)<br>Description: </code> |
255
  * Loss: [<code>TripletLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#tripletloss) with these parameters:
256
  ```json
257
  {
 
397
  ### Training Logs
398
  | Epoch | Step | Training Loss | retrieval-eval_cosine_accuracy |
399
  |:------:|:----:|:-------------:|:------------------------------:|
400
+ | 0.5 | 166 | - | 0.9731 |
401
+ | 1.0 | 332 | - | 0.9786 |
402
+ | 1.5 | 498 | - | 0.9794 |
403
+ | 1.5060 | 500 | 0.0784 | - |
404
+ | 2.0 | 664 | - | 0.9816 |
405
+ | 2.5 | 830 | - | 0.9826 |
406
+ | 3.0 | 996 | - | 0.9835 |
407
 
408
 
409
  ### Framework Versions
checkpoints/checkpoint-1328/1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 384,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
checkpoints/checkpoint-1328/README.md ADDED
@@ -0,0 +1,466 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - sentence-transformers
4
+ - sentence-similarity
5
+ - feature-extraction
6
+ - dense
7
+ - generated_from_trainer
8
+ - dataset_size:42459
9
+ - loss:TripletLoss
10
+ base_model: sentence-transformers/all-MiniLM-L6-v2
11
+ widget:
12
+ - source_sentence: policy for how can i verify if a tekton task version is still supported
13
+ by checking for the build.appstudio.redhat.com/expires-on annotation?
14
+ sentences:
15
+ - 'Helper: lib.to_array
16
+
17
+ Signature: to_array(s)
18
+
19
+ Description: '
20
+ - 'Helper: lib.pipelinerun_attestations
21
+
22
+ Signature: pipelinerun_attestations
23
+
24
+ Description: '
25
+ - 'Helper: lib.k8s.name
26
+
27
+ Signature: name(resource)
28
+
29
+ Description: '
30
+ - source_sentence: how to check attestation is missing statement field.
31
+ sentences:
32
+ - 'Helper: lib.k8s.name
33
+
34
+ Signature: name(resource)
35
+
36
+ Description: '
37
+ - 'Helper: lib.tekton.untrusted_task_refs
38
+
39
+ Signature: untrusted_task_refs(tasks)
40
+
41
+ Description: '
42
+ - 'Helper: lib.k8s.version
43
+
44
+ Signature: version(resource)
45
+
46
+ Description: '
47
+ - source_sentence: I need to ensure the operators.openshift.io/valid-subscription
48
+ annotation in the ClusterServiceVersion manifest contains a valid JSON encoded
49
+ non-empty array of strings.
50
+ sentences:
51
+ - 'Helper: lib.to_array
52
+
53
+ Signature: to_array(s)
54
+
55
+ Description: '
56
+ - 'Helper: lib.image.equal_ref
57
+
58
+ Signature: equal_ref(ref1, ref2)
59
+
60
+ Description: '
61
+ - 'Helper: lib.result_helper
62
+
63
+ Signature: result_helper(chain, failure_sprintf_params)
64
+
65
+ Description: '
66
+ - source_sentence: write a rule to deny approval for an container image with non-unique
67
+ RPM names
68
+ sentences:
69
+ - 'Helper: lib.result_helper
70
+
71
+ Signature: result_helper(chain, failure_sprintf_params)
72
+
73
+ Description: '
74
+ - 'Helper: lib.to_set
75
+
76
+ Signature: to_set(arr)
77
+
78
+ Description: '
79
+ - 'Helper: lib.rule_data_defaults
80
+
81
+ Signature: rule_data_defaults
82
+
83
+ Description: '
84
+ - source_sentence: check if i need to validate that spdx package is an operating system
85
+ component.
86
+ sentences:
87
+ - 'Helper: lib.to_set
88
+
89
+ Signature: to_set(arr)
90
+
91
+ Description: '
92
+ - 'Helper: lib.rule_data_defaults
93
+
94
+ Signature: rule_data_defaults
95
+
96
+ Description: '
97
+ - 'Helper: lib.result_helper
98
+
99
+ Signature: result_helper(chain, failure_sprintf_params)
100
+
101
+ Description: '
102
+ pipeline_tag: sentence-similarity
103
+ library_name: sentence-transformers
104
+ metrics:
105
+ - cosine_accuracy
106
+ model-index:
107
+ - name: SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
108
+ results:
109
+ - task:
110
+ type: triplet
111
+ name: Triplet
112
+ dataset:
113
+ name: retrieval eval
114
+ type: retrieval-eval
115
+ metrics:
116
+ - type: cosine_accuracy
117
+ value: 0.9834675788879395
118
+ name: Cosine Accuracy
119
+ ---
120
+
121
+ # SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
122
+
123
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
124
+
125
+ ## Model Details
126
+
127
+ ### Model Description
128
+ - **Model Type:** Sentence Transformer
129
+ - **Base model:** [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) <!-- at revision c9745ed1d9f207416be6d2e6f8de32d1f16199bf -->
130
+ - **Maximum Sequence Length:** 256 tokens
131
+ - **Output Dimensionality:** 384 dimensions
132
+ - **Similarity Function:** Cosine Similarity
133
+ <!-- - **Training Dataset:** Unknown -->
134
+ <!-- - **Language:** Unknown -->
135
+ <!-- - **License:** Unknown -->
136
+
137
+ ### Model Sources
138
+
139
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
140
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/huggingface/sentence-transformers)
141
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
142
+
143
+ ### Full Model Architecture
144
+
145
+ ```
146
+ SentenceTransformer(
147
+ (0): Transformer({'max_seq_length': 256, 'do_lower_case': False, 'architecture': 'BertModel'})
148
+ (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
149
+ (2): Normalize()
150
+ )
151
+ ```
152
+
153
+ ## Usage
154
+
155
+ ### Direct Usage (Sentence Transformers)
156
+
157
+ First install the Sentence Transformers library:
158
+
159
+ ```bash
160
+ pip install -U sentence-transformers
161
+ ```
162
+
163
+ Then you can load this model and run inference.
164
+ ```python
165
+ from sentence_transformers import SentenceTransformer
166
+
167
+ # Download from the 🤗 Hub
168
+ model = SentenceTransformer("sentence_transformers_model_id")
169
+ # Run inference
170
+ sentences = [
171
+ 'check if i need to validate that spdx package is an operating system component.',
172
+ 'Helper: lib.result_helper\nSignature: result_helper(chain, failure_sprintf_params)\nDescription: ',
173
+ 'Helper: lib.to_set\nSignature: to_set(arr)\nDescription: ',
174
+ ]
175
+ embeddings = model.encode(sentences)
176
+ print(embeddings.shape)
177
+ # [3, 384]
178
+
179
+ # Get the similarity scores for the embeddings
180
+ similarities = model.similarity(embeddings, embeddings)
181
+ print(similarities)
182
+ # tensor([[ 1.0000, 0.5582, -0.4662],
183
+ # [ 0.5582, 1.0000, -0.5014],
184
+ # [-0.4662, -0.5014, 1.0000]])
185
+ ```
186
+
187
+ <!--
188
+ ### Direct Usage (Transformers)
189
+
190
+ <details><summary>Click to see the direct usage in Transformers</summary>
191
+
192
+ </details>
193
+ -->
194
+
195
+ <!--
196
+ ### Downstream Usage (Sentence Transformers)
197
+
198
+ You can finetune this model on your own dataset.
199
+
200
+ <details><summary>Click to expand</summary>
201
+
202
+ </details>
203
+ -->
204
+
205
+ <!--
206
+ ### Out-of-Scope Use
207
+
208
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
209
+ -->
210
+
211
+ ## Evaluation
212
+
213
+ ### Metrics
214
+
215
+ #### Triplet
216
+
217
+ * Dataset: `retrieval-eval`
218
+ * Evaluated with [<code>TripletEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.TripletEvaluator)
219
+
220
+ | Metric | Value |
221
+ |:--------------------|:-----------|
222
+ | **cosine_accuracy** | **0.9835** |
223
+
224
+ <!--
225
+ ## Bias, Risks and Limitations
226
+
227
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
228
+ -->
229
+
230
+ <!--
231
+ ### Recommendations
232
+
233
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
234
+ -->
235
+
236
+ ## Training Details
237
+
238
+ ### Training Dataset
239
+
240
+ #### Unnamed Dataset
241
+
242
+ * Size: 42,459 training samples
243
+ * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
244
+ * Approximate statistics based on the first 1000 samples:
245
+ | | sentence_0 | sentence_1 | sentence_2 |
246
+ |:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
247
+ | type | string | string | string |
248
+ | details | <ul><li>min: 4 tokens</li><li>mean: 30.48 tokens</li><li>max: 159 tokens</li></ul> | <ul><li>min: 21 tokens</li><li>mean: 29.64 tokens</li><li>max: 125 tokens</li></ul> | <ul><li>min: 21 tokens</li><li>mean: 27.15 tokens</li><li>max: 125 tokens</li></ul> |
249
+ * Samples:
250
+ | sentence_0 | sentence_1 | sentence_2 |
251
+ |:-----------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------|
252
+ | <code>I need to ensure that only images from specific registries are used in our policy</code> | <code>Helper: lib.image.str<br>Signature: str(d)<br>Description: </code> | <code>Helper: lib.konflux.is_validating_image_index<br>Signature: is_validating_image_index<br>Description: </code> |
253
+ | <code>check if check warn</code> | <code>Helper: lib.tekton.expiry_of<br>Signature: expiry_of(task)<br>Description: </code> | <code>Helper: lib.tekton.untagged_task_references<br>Signature: untagged_task_references(tasks)<br>Description: </code> |
254
+ | <code>verify that task has an expiry date set.</code> | <code>Helper: lib.tekton.task_param<br>Signature: task_param(task, name)<br>Description: </code> | <code>Helper: lib.tekton.untagged_task_references<br>Signature: untagged_task_references(tasks)<br>Description: </code> |
255
+ * Loss: [<code>TripletLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#tripletloss) with these parameters:
256
+ ```json
257
+ {
258
+ "distance_metric": "TripletDistanceMetric.COSINE",
259
+ "triplet_margin": 0.5
260
+ }
261
+ ```
262
+
263
+ ### Training Hyperparameters
264
+ #### Non-Default Hyperparameters
265
+
266
+ - `eval_strategy`: steps
267
+ - `per_device_train_batch_size`: 128
268
+ - `per_device_eval_batch_size`: 128
269
+ - `num_train_epochs`: 5
270
+ - `multi_dataset_batch_sampler`: round_robin
271
+
272
+ #### All Hyperparameters
273
+ <details><summary>Click to expand</summary>
274
+
275
+ - `overwrite_output_dir`: False
276
+ - `do_predict`: False
277
+ - `eval_strategy`: steps
278
+ - `prediction_loss_only`: True
279
+ - `per_device_train_batch_size`: 128
280
+ - `per_device_eval_batch_size`: 128
281
+ - `per_gpu_train_batch_size`: None
282
+ - `per_gpu_eval_batch_size`: None
283
+ - `gradient_accumulation_steps`: 1
284
+ - `eval_accumulation_steps`: None
285
+ - `torch_empty_cache_steps`: None
286
+ - `learning_rate`: 5e-05
287
+ - `weight_decay`: 0.0
288
+ - `adam_beta1`: 0.9
289
+ - `adam_beta2`: 0.999
290
+ - `adam_epsilon`: 1e-08
291
+ - `max_grad_norm`: 1
292
+ - `num_train_epochs`: 5
293
+ - `max_steps`: -1
294
+ - `lr_scheduler_type`: linear
295
+ - `lr_scheduler_kwargs`: {}
296
+ - `warmup_ratio`: 0.0
297
+ - `warmup_steps`: 0
298
+ - `log_level`: passive
299
+ - `log_level_replica`: warning
300
+ - `log_on_each_node`: True
301
+ - `logging_nan_inf_filter`: True
302
+ - `save_safetensors`: True
303
+ - `save_on_each_node`: False
304
+ - `save_only_model`: False
305
+ - `restore_callback_states_from_checkpoint`: False
306
+ - `no_cuda`: False
307
+ - `use_cpu`: False
308
+ - `use_mps_device`: False
309
+ - `seed`: 42
310
+ - `data_seed`: None
311
+ - `jit_mode_eval`: False
312
+ - `bf16`: False
313
+ - `fp16`: False
314
+ - `fp16_opt_level`: O1
315
+ - `half_precision_backend`: auto
316
+ - `bf16_full_eval`: False
317
+ - `fp16_full_eval`: False
318
+ - `tf32`: None
319
+ - `local_rank`: 0
320
+ - `ddp_backend`: None
321
+ - `tpu_num_cores`: None
322
+ - `tpu_metrics_debug`: False
323
+ - `debug`: []
324
+ - `dataloader_drop_last`: False
325
+ - `dataloader_num_workers`: 0
326
+ - `dataloader_prefetch_factor`: None
327
+ - `past_index`: -1
328
+ - `disable_tqdm`: False
329
+ - `remove_unused_columns`: True
330
+ - `label_names`: None
331
+ - `load_best_model_at_end`: False
332
+ - `ignore_data_skip`: False
333
+ - `fsdp`: []
334
+ - `fsdp_min_num_params`: 0
335
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
336
+ - `fsdp_transformer_layer_cls_to_wrap`: None
337
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
338
+ - `parallelism_config`: None
339
+ - `deepspeed`: None
340
+ - `label_smoothing_factor`: 0.0
341
+ - `optim`: adamw_torch
342
+ - `optim_args`: None
343
+ - `adafactor`: False
344
+ - `group_by_length`: False
345
+ - `length_column_name`: length
346
+ - `project`: huggingface
347
+ - `trackio_space_id`: trackio
348
+ - `ddp_find_unused_parameters`: None
349
+ - `ddp_bucket_cap_mb`: None
350
+ - `ddp_broadcast_buffers`: False
351
+ - `dataloader_pin_memory`: True
352
+ - `dataloader_persistent_workers`: False
353
+ - `skip_memory_metrics`: True
354
+ - `use_legacy_prediction_loop`: False
355
+ - `push_to_hub`: False
356
+ - `resume_from_checkpoint`: None
357
+ - `hub_model_id`: None
358
+ - `hub_strategy`: every_save
359
+ - `hub_private_repo`: None
360
+ - `hub_always_push`: False
361
+ - `hub_revision`: None
362
+ - `gradient_checkpointing`: False
363
+ - `gradient_checkpointing_kwargs`: None
364
+ - `include_inputs_for_metrics`: False
365
+ - `include_for_metrics`: []
366
+ - `eval_do_concat_batches`: True
367
+ - `fp16_backend`: auto
368
+ - `push_to_hub_model_id`: None
369
+ - `push_to_hub_organization`: None
370
+ - `mp_parameters`:
371
+ - `auto_find_batch_size`: False
372
+ - `full_determinism`: False
373
+ - `torchdynamo`: None
374
+ - `ray_scope`: last
375
+ - `ddp_timeout`: 1800
376
+ - `torch_compile`: False
377
+ - `torch_compile_backend`: None
378
+ - `torch_compile_mode`: None
379
+ - `include_tokens_per_second`: False
380
+ - `include_num_input_tokens_seen`: no
381
+ - `neftune_noise_alpha`: None
382
+ - `optim_target_modules`: None
383
+ - `batch_eval_metrics`: False
384
+ - `eval_on_start`: False
385
+ - `use_liger_kernel`: False
386
+ - `liger_kernel_config`: None
387
+ - `eval_use_gather_object`: False
388
+ - `average_tokens_across_devices`: True
389
+ - `prompts`: None
390
+ - `batch_sampler`: batch_sampler
391
+ - `multi_dataset_batch_sampler`: round_robin
392
+ - `router_mapping`: {}
393
+ - `learning_rate_mapping`: {}
394
+
395
+ </details>
396
+
397
+ ### Training Logs
398
+ | Epoch | Step | Training Loss | retrieval-eval_cosine_accuracy |
399
+ |:------:|:----:|:-------------:|:------------------------------:|
400
+ | 0.5 | 166 | - | 0.9731 |
401
+ | 1.0 | 332 | - | 0.9786 |
402
+ | 1.5 | 498 | - | 0.9794 |
403
+ | 1.5060 | 500 | 0.0784 | - |
404
+ | 2.0 | 664 | - | 0.9816 |
405
+ | 2.5 | 830 | - | 0.9826 |
406
+ | 3.0 | 996 | - | 0.9835 |
407
+ | 3.0120 | 1000 | 0.0259 | - |
408
+ | 3.5 | 1162 | - | 0.9820 |
409
+ | 4.0 | 1328 | - | 0.9835 |
410
+
411
+
412
+ ### Framework Versions
413
+ - Python: 3.12.9
414
+ - Sentence Transformers: 5.2.0
415
+ - Transformers: 4.57.3
416
+ - PyTorch: 2.7.1+cu128
417
+ - Accelerate: 1.12.0
418
+ - Datasets: 4.4.1
419
+ - Tokenizers: 0.22.1
420
+
421
+ ## Citation
422
+
423
+ ### BibTeX
424
+
425
+ #### Sentence Transformers
426
+ ```bibtex
427
+ @inproceedings{reimers-2019-sentence-bert,
428
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
429
+ author = "Reimers, Nils and Gurevych, Iryna",
430
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
431
+ month = "11",
432
+ year = "2019",
433
+ publisher = "Association for Computational Linguistics",
434
+ url = "https://arxiv.org/abs/1908.10084",
435
+ }
436
+ ```
437
+
438
+ #### TripletLoss
439
+ ```bibtex
440
+ @misc{hermans2017defense,
441
+ title={In Defense of the Triplet Loss for Person Re-Identification},
442
+ author={Alexander Hermans and Lucas Beyer and Bastian Leibe},
443
+ year={2017},
444
+ eprint={1703.07737},
445
+ archivePrefix={arXiv},
446
+ primaryClass={cs.CV}
447
+ }
448
+ ```
449
+
450
+ <!--
451
+ ## Glossary
452
+
453
+ *Clearly define terms in order to be accessible across audiences.*
454
+ -->
455
+
456
+ <!--
457
+ ## Model Card Authors
458
+
459
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
460
+ -->
461
+
462
+ <!--
463
+ ## Model Card Contact
464
+
465
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
466
+ -->
checkpoints/checkpoint-1328/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "BertModel"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "classifier_dropout": null,
7
+ "dtype": "float32",
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 384,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 1536,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 6,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "transformers_version": "4.57.3",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 30522
25
+ }
checkpoints/checkpoint-1328/config_sentence_transformers.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "5.2.0",
4
+ "transformers": "4.57.3",
5
+ "pytorch": "2.7.1+cu128"
6
+ },
7
+ "model_type": "SentenceTransformer",
8
+ "prompts": {
9
+ "query": "",
10
+ "document": ""
11
+ },
12
+ "default_prompt_name": null,
13
+ "similarity_fn_name": "cosine"
14
+ }
checkpoints/checkpoint-1328/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2aad0732f84acdc49ba490a3ee17ce167582634b005e3d9067ccc8a75092efb0
3
+ size 90864192
checkpoints/checkpoint-1328/modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
+ ]
checkpoints/checkpoint-1328/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f3673e72ec8287ea177f115cb5663fb957efddc1b10883b179cb76674fce33c
3
+ size 180608203
checkpoints/checkpoint-1328/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ddcad12b9961e2cd9c09778b8dd5d2204823561fc5eb69ee05395a1da2d88e3
3
+ size 14645
checkpoints/checkpoint-1328/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b45bcb188fa8d46833f344fabcc21c5fbe8880314b3130c0e06c183f5891f04
3
+ size 1465
checkpoints/checkpoint-1328/sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 256,
3
+ "do_lower_case": false
4
+ }
checkpoints/checkpoint-1328/special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
checkpoints/checkpoint-1328/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoints/checkpoint-1328/tokenizer_config.json ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": false,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "extra_special_tokens": {},
49
+ "mask_token": "[MASK]",
50
+ "max_length": 128,
51
+ "model_max_length": 256,
52
+ "never_split": null,
53
+ "pad_to_multiple_of": null,
54
+ "pad_token": "[PAD]",
55
+ "pad_token_type_id": 0,
56
+ "padding_side": "right",
57
+ "sep_token": "[SEP]",
58
+ "stride": 0,
59
+ "strip_accents": null,
60
+ "tokenize_chinese_chars": true,
61
+ "tokenizer_class": "BertTokenizer",
62
+ "truncation_side": "right",
63
+ "truncation_strategy": "longest_first",
64
+ "unk_token": "[UNK]"
65
+ }
checkpoints/checkpoint-1328/trainer_state.json ADDED
@@ -0,0 +1,112 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 4.0,
6
+ "eval_steps": 166,
7
+ "global_step": 1328,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.5,
14
+ "eval_retrieval-eval_cosine_accuracy": 0.9730818271636963,
15
+ "eval_runtime": 5.0818,
16
+ "eval_samples_per_second": 0.0,
17
+ "eval_steps_per_second": 0.0,
18
+ "step": 166
19
+ },
20
+ {
21
+ "epoch": 1.0,
22
+ "eval_retrieval-eval_cosine_accuracy": 0.9785926342010498,
23
+ "eval_runtime": 5.1614,
24
+ "eval_samples_per_second": 0.0,
25
+ "eval_steps_per_second": 0.0,
26
+ "step": 332
27
+ },
28
+ {
29
+ "epoch": 1.5,
30
+ "eval_retrieval-eval_cosine_accuracy": 0.979440450668335,
31
+ "eval_runtime": 5.0153,
32
+ "eval_samples_per_second": 0.0,
33
+ "eval_steps_per_second": 0.0,
34
+ "step": 498
35
+ },
36
+ {
37
+ "epoch": 1.5060240963855422,
38
+ "grad_norm": 0.242173433303833,
39
+ "learning_rate": 1.5527199462726662e-05,
40
+ "loss": 0.0784,
41
+ "step": 500
42
+ },
43
+ {
44
+ "epoch": 2.0,
45
+ "eval_retrieval-eval_cosine_accuracy": 0.9815599918365479,
46
+ "eval_runtime": 5.097,
47
+ "eval_samples_per_second": 0.0,
48
+ "eval_steps_per_second": 0.0,
49
+ "step": 664
50
+ },
51
+ {
52
+ "epoch": 2.5,
53
+ "eval_retrieval-eval_cosine_accuracy": 0.9826197624206543,
54
+ "eval_runtime": 5.0153,
55
+ "eval_samples_per_second": 0.0,
56
+ "eval_steps_per_second": 0.0,
57
+ "step": 830
58
+ },
59
+ {
60
+ "epoch": 3.0,
61
+ "eval_retrieval-eval_cosine_accuracy": 0.9834675788879395,
62
+ "eval_runtime": 5.1638,
63
+ "eval_samples_per_second": 0.0,
64
+ "eval_steps_per_second": 0.0,
65
+ "step": 996
66
+ },
67
+ {
68
+ "epoch": 3.0120481927710845,
69
+ "grad_norm": 0.21960429847240448,
70
+ "learning_rate": 8.811282740094024e-06,
71
+ "loss": 0.0259,
72
+ "step": 1000
73
+ },
74
+ {
75
+ "epoch": 3.5,
76
+ "eval_retrieval-eval_cosine_accuracy": 0.9819839000701904,
77
+ "eval_runtime": 5.2312,
78
+ "eval_samples_per_second": 0.0,
79
+ "eval_steps_per_second": 0.0,
80
+ "step": 1162
81
+ },
82
+ {
83
+ "epoch": 4.0,
84
+ "eval_retrieval-eval_cosine_accuracy": 0.9834675788879395,
85
+ "eval_runtime": 5.0759,
86
+ "eval_samples_per_second": 0.0,
87
+ "eval_steps_per_second": 0.0,
88
+ "step": 1328
89
+ }
90
+ ],
91
+ "logging_steps": 500,
92
+ "max_steps": 1660,
93
+ "num_input_tokens_seen": 0,
94
+ "num_train_epochs": 5,
95
+ "save_steps": 332,
96
+ "stateful_callbacks": {
97
+ "TrainerControl": {
98
+ "args": {
99
+ "should_epoch_stop": false,
100
+ "should_evaluate": false,
101
+ "should_log": false,
102
+ "should_save": true,
103
+ "should_training_stop": false
104
+ },
105
+ "attributes": {}
106
+ }
107
+ },
108
+ "total_flos": 0.0,
109
+ "train_batch_size": 128,
110
+ "trial_name": null,
111
+ "trial_params": null
112
+ }
checkpoints/checkpoint-1328/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:655cb9ff00dbfe70b770f754d022211accbb66280259031bed3891ce1eb08985
3
+ size 6161
checkpoints/checkpoint-1328/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoints/checkpoint-1660/1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 384,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
checkpoints/checkpoint-1660/README.md ADDED
@@ -0,0 +1,469 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - sentence-transformers
4
+ - sentence-similarity
5
+ - feature-extraction
6
+ - dense
7
+ - generated_from_trainer
8
+ - dataset_size:42459
9
+ - loss:TripletLoss
10
+ base_model: sentence-transformers/all-MiniLM-L6-v2
11
+ widget:
12
+ - source_sentence: policy for how can i verify if a tekton task version is still supported
13
+ by checking for the build.appstudio.redhat.com/expires-on annotation?
14
+ sentences:
15
+ - 'Helper: lib.to_array
16
+
17
+ Signature: to_array(s)
18
+
19
+ Description: '
20
+ - 'Helper: lib.pipelinerun_attestations
21
+
22
+ Signature: pipelinerun_attestations
23
+
24
+ Description: '
25
+ - 'Helper: lib.k8s.name
26
+
27
+ Signature: name(resource)
28
+
29
+ Description: '
30
+ - source_sentence: how to check attestation is missing statement field.
31
+ sentences:
32
+ - 'Helper: lib.k8s.name
33
+
34
+ Signature: name(resource)
35
+
36
+ Description: '
37
+ - 'Helper: lib.tekton.untrusted_task_refs
38
+
39
+ Signature: untrusted_task_refs(tasks)
40
+
41
+ Description: '
42
+ - 'Helper: lib.k8s.version
43
+
44
+ Signature: version(resource)
45
+
46
+ Description: '
47
+ - source_sentence: I need to ensure the operators.openshift.io/valid-subscription
48
+ annotation in the ClusterServiceVersion manifest contains a valid JSON encoded
49
+ non-empty array of strings.
50
+ sentences:
51
+ - 'Helper: lib.to_array
52
+
53
+ Signature: to_array(s)
54
+
55
+ Description: '
56
+ - 'Helper: lib.image.equal_ref
57
+
58
+ Signature: equal_ref(ref1, ref2)
59
+
60
+ Description: '
61
+ - 'Helper: lib.result_helper
62
+
63
+ Signature: result_helper(chain, failure_sprintf_params)
64
+
65
+ Description: '
66
+ - source_sentence: write a rule to deny approval for an container image with non-unique
67
+ RPM names
68
+ sentences:
69
+ - 'Helper: lib.result_helper
70
+
71
+ Signature: result_helper(chain, failure_sprintf_params)
72
+
73
+ Description: '
74
+ - 'Helper: lib.to_set
75
+
76
+ Signature: to_set(arr)
77
+
78
+ Description: '
79
+ - 'Helper: lib.rule_data_defaults
80
+
81
+ Signature: rule_data_defaults
82
+
83
+ Description: '
84
+ - source_sentence: check if i need to validate that spdx package is an operating system
85
+ component.
86
+ sentences:
87
+ - 'Helper: lib.to_set
88
+
89
+ Signature: to_set(arr)
90
+
91
+ Description: '
92
+ - 'Helper: lib.rule_data_defaults
93
+
94
+ Signature: rule_data_defaults
95
+
96
+ Description: '
97
+ - 'Helper: lib.result_helper
98
+
99
+ Signature: result_helper(chain, failure_sprintf_params)
100
+
101
+ Description: '
102
+ pipeline_tag: sentence-similarity
103
+ library_name: sentence-transformers
104
+ metrics:
105
+ - cosine_accuracy
106
+ model-index:
107
+ - name: SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
108
+ results:
109
+ - task:
110
+ type: triplet
111
+ name: Triplet
112
+ dataset:
113
+ name: retrieval eval
114
+ type: retrieval-eval
115
+ metrics:
116
+ - type: cosine_accuracy
117
+ value: 0.9830436706542969
118
+ name: Cosine Accuracy
119
+ ---
120
+
121
+ # SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
122
+
123
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
124
+
125
+ ## Model Details
126
+
127
+ ### Model Description
128
+ - **Model Type:** Sentence Transformer
129
+ - **Base model:** [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) <!-- at revision c9745ed1d9f207416be6d2e6f8de32d1f16199bf -->
130
+ - **Maximum Sequence Length:** 256 tokens
131
+ - **Output Dimensionality:** 384 dimensions
132
+ - **Similarity Function:** Cosine Similarity
133
+ <!-- - **Training Dataset:** Unknown -->
134
+ <!-- - **Language:** Unknown -->
135
+ <!-- - **License:** Unknown -->
136
+
137
+ ### Model Sources
138
+
139
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
140
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/huggingface/sentence-transformers)
141
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
142
+
143
+ ### Full Model Architecture
144
+
145
+ ```
146
+ SentenceTransformer(
147
+ (0): Transformer({'max_seq_length': 256, 'do_lower_case': False, 'architecture': 'BertModel'})
148
+ (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
149
+ (2): Normalize()
150
+ )
151
+ ```
152
+
153
+ ## Usage
154
+
155
+ ### Direct Usage (Sentence Transformers)
156
+
157
+ First install the Sentence Transformers library:
158
+
159
+ ```bash
160
+ pip install -U sentence-transformers
161
+ ```
162
+
163
+ Then you can load this model and run inference.
164
+ ```python
165
+ from sentence_transformers import SentenceTransformer
166
+
167
+ # Download from the 🤗 Hub
168
+ model = SentenceTransformer("sentence_transformers_model_id")
169
+ # Run inference
170
+ sentences = [
171
+ 'check if i need to validate that spdx package is an operating system component.',
172
+ 'Helper: lib.result_helper\nSignature: result_helper(chain, failure_sprintf_params)\nDescription: ',
173
+ 'Helper: lib.to_set\nSignature: to_set(arr)\nDescription: ',
174
+ ]
175
+ embeddings = model.encode(sentences)
176
+ print(embeddings.shape)
177
+ # [3, 384]
178
+
179
+ # Get the similarity scores for the embeddings
180
+ similarities = model.similarity(embeddings, embeddings)
181
+ print(similarities)
182
+ # tensor([[ 1.0000, 0.5571, -0.4690],
183
+ # [ 0.5571, 1.0000, -0.5010],
184
+ # [-0.4690, -0.5010, 1.0000]])
185
+ ```
186
+
187
+ <!--
188
+ ### Direct Usage (Transformers)
189
+
190
+ <details><summary>Click to see the direct usage in Transformers</summary>
191
+
192
+ </details>
193
+ -->
194
+
195
+ <!--
196
+ ### Downstream Usage (Sentence Transformers)
197
+
198
+ You can finetune this model on your own dataset.
199
+
200
+ <details><summary>Click to expand</summary>
201
+
202
+ </details>
203
+ -->
204
+
205
+ <!--
206
+ ### Out-of-Scope Use
207
+
208
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
209
+ -->
210
+
211
+ ## Evaluation
212
+
213
+ ### Metrics
214
+
215
+ #### Triplet
216
+
217
+ * Dataset: `retrieval-eval`
218
+ * Evaluated with [<code>TripletEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.TripletEvaluator)
219
+
220
+ | Metric | Value |
221
+ |:--------------------|:----------|
222
+ | **cosine_accuracy** | **0.983** |
223
+
224
+ <!--
225
+ ## Bias, Risks and Limitations
226
+
227
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
228
+ -->
229
+
230
+ <!--
231
+ ### Recommendations
232
+
233
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
234
+ -->
235
+
236
+ ## Training Details
237
+
238
+ ### Training Dataset
239
+
240
+ #### Unnamed Dataset
241
+
242
+ * Size: 42,459 training samples
243
+ * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
244
+ * Approximate statistics based on the first 1000 samples:
245
+ | | sentence_0 | sentence_1 | sentence_2 |
246
+ |:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
247
+ | type | string | string | string |
248
+ | details | <ul><li>min: 4 tokens</li><li>mean: 30.48 tokens</li><li>max: 159 tokens</li></ul> | <ul><li>min: 21 tokens</li><li>mean: 29.64 tokens</li><li>max: 125 tokens</li></ul> | <ul><li>min: 21 tokens</li><li>mean: 27.15 tokens</li><li>max: 125 tokens</li></ul> |
249
+ * Samples:
250
+ | sentence_0 | sentence_1 | sentence_2 |
251
+ |:-----------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------|
252
+ | <code>I need to ensure that only images from specific registries are used in our policy</code> | <code>Helper: lib.image.str<br>Signature: str(d)<br>Description: </code> | <code>Helper: lib.konflux.is_validating_image_index<br>Signature: is_validating_image_index<br>Description: </code> |
253
+ | <code>check if check warn</code> | <code>Helper: lib.tekton.expiry_of<br>Signature: expiry_of(task)<br>Description: </code> | <code>Helper: lib.tekton.untagged_task_references<br>Signature: untagged_task_references(tasks)<br>Description: </code> |
254
+ | <code>verify that task has an expiry date set.</code> | <code>Helper: lib.tekton.task_param<br>Signature: task_param(task, name)<br>Description: </code> | <code>Helper: lib.tekton.untagged_task_references<br>Signature: untagged_task_references(tasks)<br>Description: </code> |
255
+ * Loss: [<code>TripletLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#tripletloss) with these parameters:
256
+ ```json
257
+ {
258
+ "distance_metric": "TripletDistanceMetric.COSINE",
259
+ "triplet_margin": 0.5
260
+ }
261
+ ```
262
+
263
+ ### Training Hyperparameters
264
+ #### Non-Default Hyperparameters
265
+
266
+ - `eval_strategy`: steps
267
+ - `per_device_train_batch_size`: 128
268
+ - `per_device_eval_batch_size`: 128
269
+ - `num_train_epochs`: 5
270
+ - `multi_dataset_batch_sampler`: round_robin
271
+
272
+ #### All Hyperparameters
273
+ <details><summary>Click to expand</summary>
274
+
275
+ - `overwrite_output_dir`: False
276
+ - `do_predict`: False
277
+ - `eval_strategy`: steps
278
+ - `prediction_loss_only`: True
279
+ - `per_device_train_batch_size`: 128
280
+ - `per_device_eval_batch_size`: 128
281
+ - `per_gpu_train_batch_size`: None
282
+ - `per_gpu_eval_batch_size`: None
283
+ - `gradient_accumulation_steps`: 1
284
+ - `eval_accumulation_steps`: None
285
+ - `torch_empty_cache_steps`: None
286
+ - `learning_rate`: 5e-05
287
+ - `weight_decay`: 0.0
288
+ - `adam_beta1`: 0.9
289
+ - `adam_beta2`: 0.999
290
+ - `adam_epsilon`: 1e-08
291
+ - `max_grad_norm`: 1
292
+ - `num_train_epochs`: 5
293
+ - `max_steps`: -1
294
+ - `lr_scheduler_type`: linear
295
+ - `lr_scheduler_kwargs`: {}
296
+ - `warmup_ratio`: 0.0
297
+ - `warmup_steps`: 0
298
+ - `log_level`: passive
299
+ - `log_level_replica`: warning
300
+ - `log_on_each_node`: True
301
+ - `logging_nan_inf_filter`: True
302
+ - `save_safetensors`: True
303
+ - `save_on_each_node`: False
304
+ - `save_only_model`: False
305
+ - `restore_callback_states_from_checkpoint`: False
306
+ - `no_cuda`: False
307
+ - `use_cpu`: False
308
+ - `use_mps_device`: False
309
+ - `seed`: 42
310
+ - `data_seed`: None
311
+ - `jit_mode_eval`: False
312
+ - `bf16`: False
313
+ - `fp16`: False
314
+ - `fp16_opt_level`: O1
315
+ - `half_precision_backend`: auto
316
+ - `bf16_full_eval`: False
317
+ - `fp16_full_eval`: False
318
+ - `tf32`: None
319
+ - `local_rank`: 0
320
+ - `ddp_backend`: None
321
+ - `tpu_num_cores`: None
322
+ - `tpu_metrics_debug`: False
323
+ - `debug`: []
324
+ - `dataloader_drop_last`: False
325
+ - `dataloader_num_workers`: 0
326
+ - `dataloader_prefetch_factor`: None
327
+ - `past_index`: -1
328
+ - `disable_tqdm`: False
329
+ - `remove_unused_columns`: True
330
+ - `label_names`: None
331
+ - `load_best_model_at_end`: False
332
+ - `ignore_data_skip`: False
333
+ - `fsdp`: []
334
+ - `fsdp_min_num_params`: 0
335
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
336
+ - `fsdp_transformer_layer_cls_to_wrap`: None
337
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
338
+ - `parallelism_config`: None
339
+ - `deepspeed`: None
340
+ - `label_smoothing_factor`: 0.0
341
+ - `optim`: adamw_torch
342
+ - `optim_args`: None
343
+ - `adafactor`: False
344
+ - `group_by_length`: False
345
+ - `length_column_name`: length
346
+ - `project`: huggingface
347
+ - `trackio_space_id`: trackio
348
+ - `ddp_find_unused_parameters`: None
349
+ - `ddp_bucket_cap_mb`: None
350
+ - `ddp_broadcast_buffers`: False
351
+ - `dataloader_pin_memory`: True
352
+ - `dataloader_persistent_workers`: False
353
+ - `skip_memory_metrics`: True
354
+ - `use_legacy_prediction_loop`: False
355
+ - `push_to_hub`: False
356
+ - `resume_from_checkpoint`: None
357
+ - `hub_model_id`: None
358
+ - `hub_strategy`: every_save
359
+ - `hub_private_repo`: None
360
+ - `hub_always_push`: False
361
+ - `hub_revision`: None
362
+ - `gradient_checkpointing`: False
363
+ - `gradient_checkpointing_kwargs`: None
364
+ - `include_inputs_for_metrics`: False
365
+ - `include_for_metrics`: []
366
+ - `eval_do_concat_batches`: True
367
+ - `fp16_backend`: auto
368
+ - `push_to_hub_model_id`: None
369
+ - `push_to_hub_organization`: None
370
+ - `mp_parameters`:
371
+ - `auto_find_batch_size`: False
372
+ - `full_determinism`: False
373
+ - `torchdynamo`: None
374
+ - `ray_scope`: last
375
+ - `ddp_timeout`: 1800
376
+ - `torch_compile`: False
377
+ - `torch_compile_backend`: None
378
+ - `torch_compile_mode`: None
379
+ - `include_tokens_per_second`: False
380
+ - `include_num_input_tokens_seen`: no
381
+ - `neftune_noise_alpha`: None
382
+ - `optim_target_modules`: None
383
+ - `batch_eval_metrics`: False
384
+ - `eval_on_start`: False
385
+ - `use_liger_kernel`: False
386
+ - `liger_kernel_config`: None
387
+ - `eval_use_gather_object`: False
388
+ - `average_tokens_across_devices`: True
389
+ - `prompts`: None
390
+ - `batch_sampler`: batch_sampler
391
+ - `multi_dataset_batch_sampler`: round_robin
392
+ - `router_mapping`: {}
393
+ - `learning_rate_mapping`: {}
394
+
395
+ </details>
396
+
397
+ ### Training Logs
398
+ | Epoch | Step | Training Loss | retrieval-eval_cosine_accuracy |
399
+ |:------:|:----:|:-------------:|:------------------------------:|
400
+ | 0.5 | 166 | - | 0.9731 |
401
+ | 1.0 | 332 | - | 0.9786 |
402
+ | 1.5 | 498 | - | 0.9794 |
403
+ | 1.5060 | 500 | 0.0784 | - |
404
+ | 2.0 | 664 | - | 0.9816 |
405
+ | 2.5 | 830 | - | 0.9826 |
406
+ | 3.0 | 996 | - | 0.9835 |
407
+ | 3.0120 | 1000 | 0.0259 | - |
408
+ | 3.5 | 1162 | - | 0.9820 |
409
+ | 4.0 | 1328 | - | 0.9835 |
410
+ | 4.5 | 1494 | - | 0.9835 |
411
+ | 4.5181 | 1500 | 0.0227 | - |
412
+ | 5.0 | 1660 | - | 0.9830 |
413
+
414
+
415
+ ### Framework Versions
416
+ - Python: 3.12.9
417
+ - Sentence Transformers: 5.2.0
418
+ - Transformers: 4.57.3
419
+ - PyTorch: 2.7.1+cu128
420
+ - Accelerate: 1.12.0
421
+ - Datasets: 4.4.1
422
+ - Tokenizers: 0.22.1
423
+
424
+ ## Citation
425
+
426
+ ### BibTeX
427
+
428
+ #### Sentence Transformers
429
+ ```bibtex
430
+ @inproceedings{reimers-2019-sentence-bert,
431
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
432
+ author = "Reimers, Nils and Gurevych, Iryna",
433
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
434
+ month = "11",
435
+ year = "2019",
436
+ publisher = "Association for Computational Linguistics",
437
+ url = "https://arxiv.org/abs/1908.10084",
438
+ }
439
+ ```
440
+
441
+ #### TripletLoss
442
+ ```bibtex
443
+ @misc{hermans2017defense,
444
+ title={In Defense of the Triplet Loss for Person Re-Identification},
445
+ author={Alexander Hermans and Lucas Beyer and Bastian Leibe},
446
+ year={2017},
447
+ eprint={1703.07737},
448
+ archivePrefix={arXiv},
449
+ primaryClass={cs.CV}
450
+ }
451
+ ```
452
+
453
+ <!--
454
+ ## Glossary
455
+
456
+ *Clearly define terms in order to be accessible across audiences.*
457
+ -->
458
+
459
+ <!--
460
+ ## Model Card Authors
461
+
462
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
463
+ -->
464
+
465
+ <!--
466
+ ## Model Card Contact
467
+
468
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
469
+ -->
checkpoints/checkpoint-1660/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "BertModel"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "classifier_dropout": null,
7
+ "dtype": "float32",
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 384,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 1536,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 6,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "transformers_version": "4.57.3",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 30522
25
+ }
checkpoints/checkpoint-1660/config_sentence_transformers.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "5.2.0",
4
+ "transformers": "4.57.3",
5
+ "pytorch": "2.7.1+cu128"
6
+ },
7
+ "model_type": "SentenceTransformer",
8
+ "prompts": {
9
+ "query": "",
10
+ "document": ""
11
+ },
12
+ "default_prompt_name": null,
13
+ "similarity_fn_name": "cosine"
14
+ }
checkpoints/checkpoint-1660/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d925d65a0491b220a4e1505967f7b75f4e747249498c98735ac7c97d3496ccdf
3
+ size 90864192
checkpoints/checkpoint-1660/modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
+ ]
checkpoints/checkpoint-1660/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e68e3543cf884195172d94ab018360c1137be48edbeb55c8855d2978d5e96561
3
+ size 180608203
checkpoints/checkpoint-1660/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:535a356cd0ffe920eca8a24273c362ac52b82286fb686759465904593b20bfeb
3
+ size 14645
checkpoints/checkpoint-1660/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:080157cb25d6b8140b630d6a70f028b9e651cf548599cfc118cfa1c6cace7bf0
3
+ size 1465
checkpoints/checkpoint-1660/sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 256,
3
+ "do_lower_case": false
4
+ }
checkpoints/checkpoint-1660/special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
checkpoints/checkpoint-1660/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoints/checkpoint-1660/tokenizer_config.json ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": false,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "extra_special_tokens": {},
49
+ "mask_token": "[MASK]",
50
+ "max_length": 128,
51
+ "model_max_length": 256,
52
+ "never_split": null,
53
+ "pad_to_multiple_of": null,
54
+ "pad_token": "[PAD]",
55
+ "pad_token_type_id": 0,
56
+ "padding_side": "right",
57
+ "sep_token": "[SEP]",
58
+ "stride": 0,
59
+ "strip_accents": null,
60
+ "tokenize_chinese_chars": true,
61
+ "tokenizer_class": "BertTokenizer",
62
+ "truncation_side": "right",
63
+ "truncation_strategy": "longest_first",
64
+ "unk_token": "[UNK]"
65
+ }
checkpoints/checkpoint-1660/trainer_state.json ADDED
@@ -0,0 +1,135 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 5.0,
6
+ "eval_steps": 166,
7
+ "global_step": 1660,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.5,
14
+ "eval_retrieval-eval_cosine_accuracy": 0.9730818271636963,
15
+ "eval_runtime": 5.0818,
16
+ "eval_samples_per_second": 0.0,
17
+ "eval_steps_per_second": 0.0,
18
+ "step": 166
19
+ },
20
+ {
21
+ "epoch": 1.0,
22
+ "eval_retrieval-eval_cosine_accuracy": 0.9785926342010498,
23
+ "eval_runtime": 5.1614,
24
+ "eval_samples_per_second": 0.0,
25
+ "eval_steps_per_second": 0.0,
26
+ "step": 332
27
+ },
28
+ {
29
+ "epoch": 1.5,
30
+ "eval_retrieval-eval_cosine_accuracy": 0.979440450668335,
31
+ "eval_runtime": 5.0153,
32
+ "eval_samples_per_second": 0.0,
33
+ "eval_steps_per_second": 0.0,
34
+ "step": 498
35
+ },
36
+ {
37
+ "epoch": 1.5060240963855422,
38
+ "grad_norm": 0.242173433303833,
39
+ "learning_rate": 1.5527199462726662e-05,
40
+ "loss": 0.0784,
41
+ "step": 500
42
+ },
43
+ {
44
+ "epoch": 2.0,
45
+ "eval_retrieval-eval_cosine_accuracy": 0.9815599918365479,
46
+ "eval_runtime": 5.097,
47
+ "eval_samples_per_second": 0.0,
48
+ "eval_steps_per_second": 0.0,
49
+ "step": 664
50
+ },
51
+ {
52
+ "epoch": 2.5,
53
+ "eval_retrieval-eval_cosine_accuracy": 0.9826197624206543,
54
+ "eval_runtime": 5.0153,
55
+ "eval_samples_per_second": 0.0,
56
+ "eval_steps_per_second": 0.0,
57
+ "step": 830
58
+ },
59
+ {
60
+ "epoch": 3.0,
61
+ "eval_retrieval-eval_cosine_accuracy": 0.9834675788879395,
62
+ "eval_runtime": 5.1638,
63
+ "eval_samples_per_second": 0.0,
64
+ "eval_steps_per_second": 0.0,
65
+ "step": 996
66
+ },
67
+ {
68
+ "epoch": 3.0120481927710845,
69
+ "grad_norm": 0.21960429847240448,
70
+ "learning_rate": 8.811282740094024e-06,
71
+ "loss": 0.0259,
72
+ "step": 1000
73
+ },
74
+ {
75
+ "epoch": 3.5,
76
+ "eval_retrieval-eval_cosine_accuracy": 0.9819839000701904,
77
+ "eval_runtime": 5.2312,
78
+ "eval_samples_per_second": 0.0,
79
+ "eval_steps_per_second": 0.0,
80
+ "step": 1162
81
+ },
82
+ {
83
+ "epoch": 4.0,
84
+ "eval_retrieval-eval_cosine_accuracy": 0.9834675788879395,
85
+ "eval_runtime": 5.0759,
86
+ "eval_samples_per_second": 0.0,
87
+ "eval_steps_per_second": 0.0,
88
+ "step": 1328
89
+ },
90
+ {
91
+ "epoch": 4.5,
92
+ "eval_retrieval-eval_cosine_accuracy": 0.9834675788879395,
93
+ "eval_runtime": 5.0083,
94
+ "eval_samples_per_second": 0.0,
95
+ "eval_steps_per_second": 0.0,
96
+ "step": 1494
97
+ },
98
+ {
99
+ "epoch": 4.518072289156627,
100
+ "grad_norm": 0.21867941319942474,
101
+ "learning_rate": 2.0953660174613837e-06,
102
+ "loss": 0.0227,
103
+ "step": 1500
104
+ },
105
+ {
106
+ "epoch": 5.0,
107
+ "eval_retrieval-eval_cosine_accuracy": 0.9830436706542969,
108
+ "eval_runtime": 5.0895,
109
+ "eval_samples_per_second": 0.0,
110
+ "eval_steps_per_second": 0.0,
111
+ "step": 1660
112
+ }
113
+ ],
114
+ "logging_steps": 500,
115
+ "max_steps": 1660,
116
+ "num_input_tokens_seen": 0,
117
+ "num_train_epochs": 5,
118
+ "save_steps": 332,
119
+ "stateful_callbacks": {
120
+ "TrainerControl": {
121
+ "args": {
122
+ "should_epoch_stop": false,
123
+ "should_evaluate": false,
124
+ "should_log": false,
125
+ "should_save": true,
126
+ "should_training_stop": true
127
+ },
128
+ "attributes": {}
129
+ }
130
+ },
131
+ "total_flos": 0.0,
132
+ "train_batch_size": 128,
133
+ "trial_name": null,
134
+ "trial_params": null
135
+ }
checkpoints/checkpoint-1660/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:655cb9ff00dbfe70b770f754d022211accbb66280259031bed3891ce1eb08985
3
+ size 6161
checkpoints/checkpoint-1660/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoints/eval/triplet_evaluation_retrieval-eval_results.csv CHANGED
@@ -1,26 +1,11 @@
1
  epoch,steps,accuracy_cosine
2
- 0.5,145,0.9620622396469116
3
- 1.0,290,0.9730058312416077
4
- 1.5,435,0.9764105081558228
5
- 2.0,580,0.975680947303772
6
- 2.5,725,0.9768968820571899
7
- 0.49829351535836175,146,0.9636757373809814
8
- 0.9965870307167235,292,0.9711330533027649
9
- 1.4948805460750854,438,0.9728169441223145
10
- 1.993174061433447,584,0.974500834941864
11
- 2.491467576791809,730,0.974981963634491
12
- 2.9897610921501707,876,0.9747413992881775
13
- 3.4880546075085324,1022,0.974981963634491
14
- 3.986348122866894,1168,0.974981963634491
15
- 4.484641638225256,1314,0.9757036566734314
16
- 4.982935153583618,1460,0.9757036566734314
17
- 0.5,141,0.9608283638954163
18
- 1.0,282,0.9713073968887329
19
- 1.5,423,0.9752994179725647
20
- 2.0,564,0.9750499129295349
21
- 2.5,705,0.9748004078865051
22
- 3.0,846,0.9762973785400391
23
- 3.5,987,0.9757984280586243
24
- 4.0,1128,0.976047933101654
25
- 4.5,1269,0.9750499129295349
26
- 5.0,1410,0.9750499129295349
 
1
  epoch,steps,accuracy_cosine
2
+ 0.5,166,0.9730818271636963
3
+ 1.0,332,0.9785926342010498
4
+ 1.5,498,0.979440450668335
5
+ 2.0,664,0.9815599918365479
6
+ 2.5,830,0.9826197624206543
7
+ 3.0,996,0.9834675788879395
8
+ 3.5,1162,0.9819839000701904
9
+ 4.0,1328,0.9834675788879395
10
+ 4.5,1494,0.9834675788879395
11
+ 5.0,1660,0.9830436706542969
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoints/runs/Dec13_16-18-12_rego-trainer-0/events.out.tfevents.1765642693.rego-trainer-0.4819.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b738af4d9ec875eecde1711029d777c2abb8cab0931f90d79f533076b9dc38f6
3
+ size 8447
eval/triplet_evaluation_retrieval-eval_results.csv CHANGED
@@ -1,13 +1,6 @@
1
  epoch,steps,accuracy_cosine
2
- 1.0,290,0.9730058312416077
3
- 2.0,580,0.975680947303772
4
- 1.0,293,0.9718546867370605
5
- 2.0,586,0.974500834941864
6
- 3.0,879,0.974500834941864
7
- 4.0,1172,0.9754630923271179
8
- 5.0,1465,0.9757036566734314
9
- 1.0,282,0.9713073968887329
10
- 2.0,564,0.9750499129295349
11
- 3.0,846,0.9762973785400391
12
- 4.0,1128,0.976047933101654
13
- 5.0,1410,0.9750499129295349
 
1
  epoch,steps,accuracy_cosine
2
+ 1.0,332,0.9785926342010498
3
+ 2.0,664,0.9815599918365479
4
+ 3.0,996,0.9834675788879395
5
+ 4.0,1328,0.9834675788879395
6
+ 5.0,1660,0.9830436706542969
 
 
 
 
 
 
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:28fbf416ad4f25af4d2d832531c5941855143597d18b824cd223fbea1ebd6434
3
  size 90864192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:745aba2b4a7f3a5bbd9c3e086d962d7e047f04db05a4ed6d52f7240baf135d67
3
  size 90864192
training_info.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "base_model": "sentence-transformers/all-MiniLM-L6-v2",
3
- "train_examples": 36065,
4
- "eval_examples": 4008,
5
  "epochs": 5,
6
  "batch_size": 128,
7
  "learning_rate": 2e-05,
 
1
  {
2
  "base_model": "sentence-transformers/all-MiniLM-L6-v2",
3
+ "train_examples": 42459,
4
+ "eval_examples": 4718,
5
  "epochs": 5,
6
  "batch_size": 128,
7
  "learning_rate": 2e-05,