Datasets:

Modalities:
Text
Formats:
parquet
Size:
< 1K
Libraries:
Datasets
pandas
PabloAccuosto commited on
Commit
ad7927c
·
verified ·
1 Parent(s): ee4ab34

Upload folder using huggingface_hub

Browse files
Files changed (4) hide show
  1. .argilla/dataset.json +16 -0
  2. .argilla/settings.json +170 -0
  3. .argilla/version.json +3 -0
  4. README.md +176 -70
.argilla/dataset.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "id": "53a2c2fe-9a3f-4245-aa35-9bcc061876d9",
3
+ "name": "scilake-ccam",
4
+ "guidelines": "# Validation guidelines for CCAM entities\n## Task Description\nYour task is to validate the extraction of the different entities and their linking to their closest matching entries in the vocabulary created for SciLake.\n\n## What to Validate\nFor each record, please verify the following:\n1. **Entity Spans:** Are all text spans correctly identified? Are the span boundaries accurate?\n2. **Entity Types:** Are entity types correctly assigned?\n3. **Entity Linking:** Are the matching entities in the vocabulary correctly assigned?\n\n## Instructions\n1. Carefully read the texts.\n2. Review the NER spans and correct them if:\n- The boundaries (start/end) are incorrect\n- The entity label is wrong\n3. Verify that the extracted entities are correctly linked to their closest match in the vocabulary\n4. Add any comments or feedback you deem relevant\n\n## Validation Guidelines\n- Entity Annotations: Mark spans as \"Correct\" only if boundaries and labels are accurate.\n- Entity Extraction: Mark as \"Correct\" if all energy (storage) types mentioned are extracted; \"Partially correct\" if any are missing or incorrect.\n- Vocabulary Linking: Mark as \"Correct\" if all links are to the appropriate entries. Use \"Partially correct\" if any are incorrect.\n\n## Entities\n- `communicationType`: the technology used for communication (eg. 4G, 5G), NOT who is connecting with whom\n- `sensorType`: the type of sensor (eg. camera, LIDAR)\n- `scenarioType`: the driving scenario (eg. cut in, lane keeping)\n- `vehicleType`: the type of vehicle (eg. car, truck)\n- `VRUType`: vulnerable road users (eg. pedestrian, cyclist)\n- `entityConnectionType`: type of connection between entities (eg. V2V, V2I), NOT the technology\n- `levelOfAutomation`: entities related to automation (eg. ALKS, driver assistance) and their relation to the FAME level of automation",
5
+ "allow_extra_metadata": false,
6
+ "status": "ready",
7
+ "distribution": {
8
+ "strategy": "overlap",
9
+ "min_submitted": 1
10
+ },
11
+ "metadata": null,
12
+ "workspace_id": "0756eadb-468f-4c06-88c4-51a3fa6f665f",
13
+ "last_activity_at": "2025-06-19T09:38:08.565572",
14
+ "inserted_at": "2025-04-09T13:03:19.498269",
15
+ "updated_at": "2025-04-09T13:03:20.918402"
16
+ }
.argilla/settings.json ADDED
@@ -0,0 +1,170 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "guidelines": "# Validation guidelines for CCAM entities\n## Task Description\nYour task is to validate the extraction of the different entities and their linking to their closest matching entries in the vocabulary created for SciLake.\n\n## What to Validate\nFor each record, please verify the following:\n1. **Entity Spans:** Are all text spans correctly identified? Are the span boundaries accurate?\n2. **Entity Types:** Are entity types correctly assigned?\n3. **Entity Linking:** Are the matching entities in the vocabulary correctly assigned?\n\n## Instructions\n1. Carefully read the texts.\n2. Review the NER spans and correct them if:\n- The boundaries (start/end) are incorrect\n- The entity label is wrong\n3. Verify that the extracted entities are correctly linked to their closest match in the vocabulary\n4. Add any comments or feedback you deem relevant\n\n## Validation Guidelines\n- Entity Annotations: Mark spans as \"Correct\" only if boundaries and labels are accurate.\n- Entity Extraction: Mark as \"Correct\" if all energy (storage) types mentioned are extracted; \"Partially correct\" if any are missing or incorrect.\n- Vocabulary Linking: Mark as \"Correct\" if all links are to the appropriate entries. Use \"Partially correct\" if any are incorrect.\n\n## Entities\n- `communicationType`: the technology used for communication (eg. 4G, 5G), NOT who is connecting with whom\n- `sensorType`: the type of sensor (eg. camera, LIDAR)\n- `scenarioType`: the driving scenario (eg. cut in, lane keeping)\n- `vehicleType`: the type of vehicle (eg. car, truck)\n- `VRUType`: vulnerable road users (eg. pedestrian, cyclist)\n- `entityConnectionType`: type of connection between entities (eg. V2V, V2I), NOT the technology\n- `levelOfAutomation`: entities related to automation (eg. ALKS, driver assistance) and their relation to the FAME level of automation",
3
+ "allow_extra_metadata": false,
4
+ "distribution": {
5
+ "strategy": "overlap",
6
+ "min_submitted": 1
7
+ },
8
+ "fields": [
9
+ {
10
+ "id": "269959b7-78e9-4fce-8dec-f37ef53a65a6",
11
+ "name": "text",
12
+ "title": "Text",
13
+ "required": true,
14
+ "settings": {
15
+ "type": "text",
16
+ "use_markdown": false
17
+ },
18
+ "dataset_id": "53a2c2fe-9a3f-4245-aa35-9bcc061876d9",
19
+ "inserted_at": "2025-04-09T13:03:19.759558",
20
+ "updated_at": "2025-04-09T13:03:19.759558"
21
+ },
22
+ {
23
+ "id": "bdc682f9-1e61-4144-90ee-8dbfe0e07cb3",
24
+ "name": "links",
25
+ "title": "Linked entities",
26
+ "required": true,
27
+ "settings": {
28
+ "type": "text",
29
+ "use_markdown": true
30
+ },
31
+ "dataset_id": "53a2c2fe-9a3f-4245-aa35-9bcc061876d9",
32
+ "inserted_at": "2025-04-09T13:03:19.863417",
33
+ "updated_at": "2025-04-09T13:03:19.863417"
34
+ }
35
+ ],
36
+ "questions": [
37
+ {
38
+ "id": "7c476ed4-1293-46b2-afcc-a9c9339e5e85",
39
+ "name": "span_label",
40
+ "title": "Select and classify the tokens according to the specified categories.",
41
+ "description": null,
42
+ "required": true,
43
+ "settings": {
44
+ "type": "span",
45
+ "field": "text",
46
+ "options": [
47
+ {
48
+ "value": "communicationType",
49
+ "text": "communicationType",
50
+ "description": null
51
+ },
52
+ {
53
+ "value": "sensorType",
54
+ "text": "sensorType",
55
+ "description": null
56
+ },
57
+ {
58
+ "value": "scenarioType",
59
+ "text": "scenarioType",
60
+ "description": null
61
+ },
62
+ {
63
+ "value": "vehicleType",
64
+ "text": "vehicleType",
65
+ "description": null
66
+ },
67
+ {
68
+ "value": "VRUType",
69
+ "text": "VRUType",
70
+ "description": null
71
+ },
72
+ {
73
+ "value": "entityConnectionType",
74
+ "text": "entityConnectionType",
75
+ "description": null
76
+ },
77
+ {
78
+ "value": "levelOfAutomation",
79
+ "text": "levelOfAutomation",
80
+ "description": null
81
+ }
82
+ ],
83
+ "visible_options": 7,
84
+ "allow_overlapping": true,
85
+ "allow_character_annotation": true
86
+ },
87
+ "dataset_id": "53a2c2fe-9a3f-4245-aa35-9bcc061876d9",
88
+ "inserted_at": "2025-04-09T13:03:20.046436",
89
+ "updated_at": "2025-04-09T13:03:20.046436"
90
+ },
91
+ {
92
+ "id": "94a5c85a-7b84-4137-8191-6f08c2742ab2",
93
+ "name": "assess_ner",
94
+ "title": "Extracted entity validation",
95
+ "description": "Are the extracted entities correct?",
96
+ "required": true,
97
+ "settings": {
98
+ "type": "label_selection",
99
+ "options": [
100
+ {
101
+ "value": "Correct",
102
+ "text": "Correct",
103
+ "description": null
104
+ },
105
+ {
106
+ "value": "Partially correct",
107
+ "text": "Partially correct",
108
+ "description": null
109
+ },
110
+ {
111
+ "value": "Incorrect",
112
+ "text": "Incorrect",
113
+ "description": null
114
+ }
115
+ ],
116
+ "visible_options": 3
117
+ },
118
+ "dataset_id": "53a2c2fe-9a3f-4245-aa35-9bcc061876d9",
119
+ "inserted_at": "2025-04-09T13:03:20.241002",
120
+ "updated_at": "2025-04-09T13:03:20.241002"
121
+ },
122
+ {
123
+ "id": "c14d6665-ab0e-4725-955e-142d0c2bb3ea",
124
+ "name": "assess_nel",
125
+ "title": "Linked vocabulary entity validation",
126
+ "description": "Are the linked entities in the vocabulary correct?",
127
+ "required": true,
128
+ "settings": {
129
+ "type": "label_selection",
130
+ "options": [
131
+ {
132
+ "value": "Correct",
133
+ "text": "Correct",
134
+ "description": null
135
+ },
136
+ {
137
+ "value": "Partially correct",
138
+ "text": "Partially correct",
139
+ "description": null
140
+ },
141
+ {
142
+ "value": "Incorrect",
143
+ "text": "Incorrect",
144
+ "description": null
145
+ }
146
+ ],
147
+ "visible_options": 3
148
+ },
149
+ "dataset_id": "53a2c2fe-9a3f-4245-aa35-9bcc061876d9",
150
+ "inserted_at": "2025-04-09T13:03:20.508913",
151
+ "updated_at": "2025-04-09T13:03:20.508913"
152
+ },
153
+ {
154
+ "id": "74893055-67ca-432d-8dfb-bd43fbaec3a3",
155
+ "name": "comments",
156
+ "title": "Comments",
157
+ "description": "Additional comments",
158
+ "required": false,
159
+ "settings": {
160
+ "type": "text",
161
+ "use_markdown": false
162
+ },
163
+ "dataset_id": "53a2c2fe-9a3f-4245-aa35-9bcc061876d9",
164
+ "inserted_at": "2025-04-09T13:03:20.719390",
165
+ "updated_at": "2025-04-09T13:03:20.719390"
166
+ }
167
+ ],
168
+ "metadata": [],
169
+ "vectors": []
170
+ }
.argilla/version.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "argilla": "2.6.0"
3
+ }
README.md CHANGED
@@ -1,72 +1,178 @@
1
  ---
2
- dataset_info:
3
- features:
4
- - name: id
5
- dtype: string
6
- - name: status
7
- dtype: string
8
- - name: inserted_at
9
- dtype: timestamp[us]
10
- - name: updated_at
11
- dtype: timestamp[us]
12
- - name: _server_id
13
- dtype: string
14
- - name: text
15
- dtype: string
16
- - name: links
17
- dtype: string
18
- - name: span_label.responses
19
- list:
20
- list:
21
- - name: end
22
- dtype: int64
23
- - name: label
24
- dtype: string
25
- - name: start
26
- dtype: int64
27
- - name: span_label.responses.users
28
- sequence: string
29
- - name: span_label.responses.status
30
- sequence: string
31
- - name: assess_ner.responses
32
- sequence: string
33
- - name: assess_ner.responses.users
34
- sequence: string
35
- - name: assess_ner.responses.status
36
- sequence: string
37
- - name: assess_nel.responses
38
- sequence: string
39
- - name: assess_nel.responses.users
40
- sequence: string
41
- - name: assess_nel.responses.status
42
- sequence: string
43
- - name: comments.responses
44
- sequence: string
45
- - name: comments.responses.users
46
- sequence: string
47
- - name: comments.responses.status
48
- sequence: string
49
- - name: span_label.suggestion
50
- list:
51
- - name: end
52
- dtype: int64
53
- - name: label
54
- dtype: string
55
- - name: start
56
- dtype: int64
57
- - name: span_label.suggestion.agent
58
- dtype: 'null'
59
- - name: span_label.suggestion.score
60
- dtype: 'null'
61
- splits:
62
- - name: train
63
- num_bytes: 525354
64
- num_examples: 191
65
- download_size: 249330
66
- dataset_size: 525354
67
- configs:
68
- - config_name: default
69
- data_files:
70
- - split: train
71
- path: data/train-*
72
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ tags:
3
+ - rlfh
4
+ - argilla
5
+ - human-feedback
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  ---
7
+
8
+ # Dataset Card for scilake-ccam
9
+
10
+
11
+
12
+
13
+
14
+
15
+
16
+ This dataset has been created with [Argilla](https://github.com/argilla-io/argilla). As shown in the sections below, this dataset can be loaded into your Argilla server as explained in [Load with Argilla](#load-with-argilla), or used directly with the `datasets` library in [Load with `datasets`](#load-with-datasets).
17
+
18
+
19
+ ## Using this dataset with Argilla
20
+
21
+ To load with Argilla, you'll just need to install Argilla as `pip install argilla --upgrade` and then use the following code:
22
+
23
+ ```python
24
+ import argilla as rg
25
+
26
+ ds = rg.Dataset.from_hub("SIRIS-Lab/scilake-ccam", settings="auto")
27
+ ```
28
+
29
+ This will load the settings and records from the dataset repository and push them to you Argilla server for exploration and annotation.
30
+
31
+ ## Using this dataset with `datasets`
32
+
33
+ To load the records of this dataset with `datasets`, you'll just need to install `datasets` as `pip install datasets --upgrade` and then use the following code:
34
+
35
+ ```python
36
+ from datasets import load_dataset
37
+
38
+ ds = load_dataset("SIRIS-Lab/scilake-ccam")
39
+ ```
40
+
41
+ This will only load the records of the dataset, but not the Argilla settings.
42
+
43
+ ## Dataset Structure
44
+
45
+ This dataset repo contains:
46
+
47
+ * Dataset records in a format compatible with HuggingFace `datasets`. These records will be loaded automatically when using `rg.Dataset.from_hub` and can be loaded independently using the `datasets` library via `load_dataset`.
48
+ * The [annotation guidelines](#annotation-guidelines) that have been used for building and curating the dataset, if they've been defined in Argilla.
49
+ * A dataset configuration folder conforming to the Argilla dataset format in `.argilla`.
50
+
51
+ The dataset is created in Argilla with: **fields**, **questions**, **suggestions**, **metadata**, **vectors**, and **guidelines**.
52
+
53
+ ### Fields
54
+
55
+ The **fields** are the features or text of a dataset's records. For example, the 'text' column of a text classification dataset of the 'prompt' column of an instruction following dataset.
56
+
57
+ | Field Name | Title | Type | Required |
58
+ | ---------- | ----- | ---- | -------- |
59
+ | text | Text | text | True |
60
+ | links | Linked entities | text | True |
61
+
62
+
63
+ ### Questions
64
+
65
+ The **questions** are the questions that will be asked to the annotators. They can be of different types, such as rating, text, label_selection, multi_label_selection, or ranking.
66
+
67
+ | Question Name | Title | Type | Required | Description | Values/Labels |
68
+ | ------------- | ----- | ---- | -------- | ----------- | ------------- |
69
+ | span_label | Select and classify the tokens according to the specified categories. | span | True | N/A | ['communicationType', 'sensorType', 'scenarioType', 'vehicleType', 'VRUType', 'entityConnectionType', 'levelOfAutomation'] |
70
+ | assess_ner | Extracted entity validation | label_selection | True | Are the extracted entities correct? | ['Correct', 'Partially correct', 'Incorrect'] |
71
+ | assess_nel | Linked vocabulary entity validation | label_selection | True | Are the linked entities in the vocabulary correct? | ['Correct', 'Partially correct', 'Incorrect'] |
72
+ | comments | Comments | text | False | Additional comments | N/A |
73
+
74
+
75
+ <!-- check length of metadata properties -->
76
+
77
+
78
+
79
+
80
+ ### Data Splits
81
+
82
+ The dataset contains a single split, which is `train`.
83
+
84
+ ## Dataset Creation
85
+
86
+ ### Curation Rationale
87
+
88
+ [More Information Needed]
89
+
90
+ ### Source Data
91
+
92
+ #### Initial Data Collection and Normalization
93
+
94
+ [More Information Needed]
95
+
96
+ #### Who are the source language producers?
97
+
98
+ [More Information Needed]
99
+
100
+ ### Annotations
101
+
102
+ #### Annotation guidelines
103
+
104
+ # Validation guidelines for CCAM entities
105
+ ## Task Description
106
+ Your task is to validate the extraction of the different entities and their linking to their closest matching entries in the vocabulary created for SciLake.
107
+
108
+ ## What to Validate
109
+ For each record, please verify the following:
110
+ 1. **Entity Spans:** Are all text spans correctly identified? Are the span boundaries accurate?
111
+ 2. **Entity Types:** Are entity types correctly assigned?
112
+ 3. **Entity Linking:** Are the matching entities in the vocabulary correctly assigned?
113
+
114
+ ## Instructions
115
+ 1. Carefully read the texts.
116
+ 2. Review the NER spans and correct them if:
117
+ - The boundaries (start/end) are incorrect
118
+ - The entity label is wrong
119
+ 3. Verify that the extracted entities are correctly linked to their closest match in the vocabulary
120
+ 4. Add any comments or feedback you deem relevant
121
+
122
+ ## Validation Guidelines
123
+ - Entity Annotations: Mark spans as "Correct" only if boundaries and labels are accurate.
124
+ - Entity Extraction: Mark as "Correct" if all energy (storage) types mentioned are extracted; "Partially correct" if any are missing or incorrect.
125
+ - Vocabulary Linking: Mark as "Correct" if all links are to the appropriate entries. Use "Partially correct" if any are incorrect.
126
+
127
+ ## Entities
128
+ - `communicationType`: the technology used for communication (eg. 4G, 5G), NOT who is connecting with whom
129
+ - `sensorType`: the type of sensor (eg. camera, LIDAR)
130
+ - `scenarioType`: the driving scenario (eg. cut in, lane keeping)
131
+ - `vehicleType`: the type of vehicle (eg. car, truck)
132
+ - `VRUType`: vulnerable road users (eg. pedestrian, cyclist)
133
+ - `entityConnectionType`: type of connection between entities (eg. V2V, V2I), NOT the technology
134
+ - `levelOfAutomation`: entities related to automation (eg. ALKS, driver assistance) and their relation to the FAME level of automation
135
+
136
+ #### Annotation process
137
+
138
+ [More Information Needed]
139
+
140
+ #### Who are the annotators?
141
+
142
+ [More Information Needed]
143
+
144
+ ### Personal and Sensitive Information
145
+
146
+ [More Information Needed]
147
+
148
+ ## Considerations for Using the Data
149
+
150
+ ### Social Impact of Dataset
151
+
152
+ [More Information Needed]
153
+
154
+ ### Discussion of Biases
155
+
156
+ [More Information Needed]
157
+
158
+ ### Other Known Limitations
159
+
160
+ [More Information Needed]
161
+
162
+ ## Additional Information
163
+
164
+ ### Dataset Curators
165
+
166
+ [More Information Needed]
167
+
168
+ ### Licensing Information
169
+
170
+ [More Information Needed]
171
+
172
+ ### Citation Information
173
+
174
+ [More Information Needed]
175
+
176
+ ### Contributions
177
+
178
+ [More Information Needed]