RiccardoDav commited on
Commit
ec076f3
·
verified ·
1 Parent(s): d96260c

Dear model owner(s),
We are a group of researchers investigating the usefulness of sharing AIBOMs (Artificial Intelligence Bill of Materials) to document AI models – AIBOMs are machine-readable structured lists of components (e.g., datasets and models) used to enhance transparency in AI-model supply chains.

To pursue the above-mentioned objective, we identified popular models on HuggingFace and, based on your model card (and some configuration information available in HuggingFace), we generated your AIBOM according to the CyclonDX (v1.6) standard (see https://cyclonedx.org/docs/1.6/json/). AIBOMs are generated as JSON files by using the following open-source supporting tool: https://github.com/MSR4SBOM/ALOHA (technical details are available in the research paper: https://github.com/MSR4SBOM/ALOHA/blob/main/ALOHA.pdf).

The JSON file in this pull request is your AIBOM (see https://github.com/MSR4SBOM/ALOHA/blob/main/documentation.json for details on its structure).

Clearly, the submitted AIBOM matches the current model information, yet it can be easily regenerated when the model evolves, using the aforementioned AIBOM generator tool.

We open this pull request containing an AIBOM of your AI model, and hope it will be considered. We would also like to hear your opinion on the usefulness (or not) of AIBOM by answering a 3-minute anonymous survey: https://forms.gle/WGffSQD5dLoWttEe7.

Thanks in advance, and regards,
Riccardo D’Avino, Fatima Ahmed, Sabato Nocera, Simone Romano, Giuseppe Scanniello (University of Salerno, Italy),
Massimiliano Di Penta (University of Sannio, Italy),
The MSR4SBOM team

cognitivecomputations_Dolphin3.0-R1-Mistral-24B.json ADDED
@@ -0,0 +1,812 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bomFormat": "CycloneDX",
3
+ "specVersion": "1.6",
4
+ "serialNumber": "urn:uuid:e6ef2cd0-b2c1-4b2e-ba7c-adb01c441786",
5
+ "version": 1,
6
+ "metadata": {
7
+ "timestamp": "2025-06-05T09:34:41.455909+00:00",
8
+ "component": {
9
+ "type": "machine-learning-model",
10
+ "bom-ref": "cognitivecomputations/Dolphin3.0-R1-Mistral-24B-6784ad22-545f-555e-aac7-abd08a7c4fc0",
11
+ "name": "cognitivecomputations/Dolphin3.0-R1-Mistral-24B",
12
+ "externalReferences": [
13
+ {
14
+ "url": "https://huggingface.co/cognitivecomputations/Dolphin3.0-R1-Mistral-24B",
15
+ "type": "documentation"
16
+ }
17
+ ],
18
+ "modelCard": {
19
+ "modelParameters": {
20
+ "task": "text-generation",
21
+ "architectureFamily": "mistral",
22
+ "modelArchitecture": "MistralForCausalLM",
23
+ "datasets": [
24
+ {
25
+ "ref": "cognitivecomputations/dolphin-r1-e86781ae-297a-5171-8a0d-61460dd40738"
26
+ },
27
+ {
28
+ "ref": "OpenCoder-LLM/opc-sft-stage1-f0572249-17d5-5847-8e5a-72583934eca0"
29
+ },
30
+ {
31
+ "ref": "OpenCoder-LLM/opc-sft-stage2-d7e9795c-e343-5ebc-b785-9718e7d737e8"
32
+ },
33
+ {
34
+ "ref": "microsoft/orca-agentinstruct-1M-v1-bbc92138-5aa0-5737-8ce6-93043b04b4dd"
35
+ },
36
+ {
37
+ "ref": "microsoft/orca-math-word-problems-200k-611afa9f-b6db-5b9f-9a51-598e4ce79d0e"
38
+ },
39
+ {
40
+ "ref": "NousResearch/hermes-function-calling-v1-a6d53a4d-e191-5d88-867f-4472e0bdc9f6"
41
+ },
42
+ {
43
+ "ref": "AI-MO/NuminaMath-CoT-3aa976f5-9ca4-5435-8542-1a123856aafb"
44
+ },
45
+ {
46
+ "ref": "AI-MO/NuminaMath-TIR-7930ec33-be02-5b4c-9fd0-00effafbc6ce"
47
+ },
48
+ {
49
+ "ref": "allenai/tulu-3-sft-mixture-4f86da52-fbf0-52de-9b77-716bafb7e098"
50
+ },
51
+ {
52
+ "ref": "cognitivecomputations/dolphin-coder-69688d29-ae99-5d6e-828c-cfc37b7221b1"
53
+ },
54
+ {
55
+ "ref": "HuggingFaceTB/smoltalk-0cc0e162-0f38-50bd-b5b0-169bcd97515b"
56
+ },
57
+ {
58
+ "ref": "cognitivecomputations/samantha-data-9c52f41f-feb0-51dd-921a-2a581d9f2fc7"
59
+ },
60
+ {
61
+ "ref": "m-a-p/CodeFeedback-Filtered-Instruction-9012249d-db87-5b91-a4e9-2e2bd74e6053"
62
+ },
63
+ {
64
+ "ref": "m-a-p/Code-Feedback-f4d189a1-046d-5a43-8007-a1eec34e9a7f"
65
+ }
66
+ ]
67
+ },
68
+ "properties": [
69
+ {
70
+ "name": "library_name",
71
+ "value": "transformers"
72
+ },
73
+ {
74
+ "name": "base_model",
75
+ "value": "mistralai/Mistral-Small-24B-Base-2501"
76
+ }
77
+ ]
78
+ },
79
+ "authors": [
80
+ {
81
+ "name": "cognitivecomputations"
82
+ }
83
+ ],
84
+ "tags": [
85
+ "transformers",
86
+ "safetensors",
87
+ "mistral",
88
+ "text-generation",
89
+ "conversational",
90
+ "en",
91
+ "dataset:cognitivecomputations/dolphin-r1",
92
+ "dataset:OpenCoder-LLM/opc-sft-stage1",
93
+ "dataset:OpenCoder-LLM/opc-sft-stage2",
94
+ "dataset:microsoft/orca-agentinstruct-1M-v1",
95
+ "dataset:microsoft/orca-math-word-problems-200k",
96
+ "dataset:NousResearch/hermes-function-calling-v1",
97
+ "dataset:AI-MO/NuminaMath-CoT",
98
+ "dataset:AI-MO/NuminaMath-TIR",
99
+ "dataset:allenai/tulu-3-sft-mixture",
100
+ "dataset:cognitivecomputations/dolphin-coder",
101
+ "dataset:HuggingFaceTB/smoltalk",
102
+ "dataset:cognitivecomputations/samantha-data",
103
+ "dataset:m-a-p/CodeFeedback-Filtered-Instruction",
104
+ "dataset:m-a-p/Code-Feedback",
105
+ "base_model:mistralai/Mistral-Small-24B-Base-2501",
106
+ "base_model:finetune:mistralai/Mistral-Small-24B-Base-2501",
107
+ "autotrain_compatible",
108
+ "text-generation-inference",
109
+ "endpoints_compatible",
110
+ "region:us"
111
+ ]
112
+ }
113
+ },
114
+ "components": [
115
+ {
116
+ "type": "data",
117
+ "bom-ref": "cognitivecomputations/dolphin-r1-e86781ae-297a-5171-8a0d-61460dd40738",
118
+ "name": "cognitivecomputations/dolphin-r1",
119
+ "data": [
120
+ {
121
+ "type": "dataset",
122
+ "bom-ref": "cognitivecomputations/dolphin-r1-e86781ae-297a-5171-8a0d-61460dd40738",
123
+ "name": "cognitivecomputations/dolphin-r1",
124
+ "contents": {
125
+ "url": "https://huggingface.co/datasets/cognitivecomputations/dolphin-r1",
126
+ "properties": [
127
+ {
128
+ "name": "configs",
129
+ "value": "Name of the dataset subset: nonreasoning {\"split\": \"train\", \"path\": \"dolphin-r1-nonreasoning.jsonl\"}"
130
+ },
131
+ {
132
+ "name": "configs",
133
+ "value": "Name of the dataset subset: reasoning-deepseek {\"split\": \"train\", \"path\": \"dolphin-r1-reasoning-deepseek.jsonl\"}"
134
+ },
135
+ {
136
+ "name": "configs",
137
+ "value": "Name of the dataset subset: reasoning-flash {\"split\": \"train\", \"path\": \"dolphin-r1-reasoning-flash.jsonl\"}"
138
+ },
139
+ {
140
+ "name": "license",
141
+ "value": "apache-2.0"
142
+ }
143
+ ]
144
+ },
145
+ "governance": {
146
+ "owners": [
147
+ {
148
+ "organization": {
149
+ "name": "cognitivecomputations",
150
+ "url": "https://huggingface.co/cognitivecomputations"
151
+ }
152
+ }
153
+ ]
154
+ },
155
+ "description": "\n\t\n\t\t\n\t\tDolphin R1 \ud83d\udc2c\n\t\n\nAn Apache-2.0 dataset curated by Eric Hartford and Cognitive Computations\n\nDiscord: https://discord.gg/cognitivecomputations\n\n\n\n\t\n\t\t\n\t\n\t\n\t\tSponsors\n\t\n\nOur appreciation for the generous sponsors of Dolphin R1 - Without whom this dataset could not exist.\n\nDria https://x.com/driaforall - Inference Sponsor (DeepSeek)\nChutes https://x.com/rayon_labs - Inference Sponsor (Flash)\nCrusoe Cloud - Compute Sponsor\nAndreessen Horowitz - provided the grant that originally launched\u2026 See the full description on the dataset page: https://huggingface.co/datasets/cognitivecomputations/dolphin-r1."
156
+ }
157
+ ]
158
+ },
159
+ {
160
+ "type": "data",
161
+ "bom-ref": "OpenCoder-LLM/opc-sft-stage1-f0572249-17d5-5847-8e5a-72583934eca0",
162
+ "name": "OpenCoder-LLM/opc-sft-stage1",
163
+ "data": [
164
+ {
165
+ "type": "dataset",
166
+ "bom-ref": "OpenCoder-LLM/opc-sft-stage1-f0572249-17d5-5847-8e5a-72583934eca0",
167
+ "name": "OpenCoder-LLM/opc-sft-stage1",
168
+ "contents": {
169
+ "url": "https://huggingface.co/datasets/OpenCoder-LLM/opc-sft-stage1",
170
+ "properties": [
171
+ {
172
+ "name": "configs",
173
+ "value": "Name of the dataset subset: filtered_infinity_instruct {\"split\": \"train\", \"path\": \"data/filtered_infinity_instruct-*\"}"
174
+ },
175
+ {
176
+ "name": "configs",
177
+ "value": "Name of the dataset subset: largescale_diverse_instruct {\"split\": \"train\", \"path\": \"data/largescale_diverse_instruct-*\"}"
178
+ },
179
+ {
180
+ "name": "configs",
181
+ "value": "Name of the dataset subset: realuser_instruct {\"split\": \"train\", \"path\": \"data/realuser_instruct-*\"}"
182
+ },
183
+ {
184
+ "name": "license",
185
+ "value": "mit"
186
+ }
187
+ ]
188
+ },
189
+ "governance": {
190
+ "owners": [
191
+ {
192
+ "organization": {
193
+ "name": "OpenCoder-LLM",
194
+ "url": "https://huggingface.co/OpenCoder-LLM"
195
+ }
196
+ }
197
+ ]
198
+ },
199
+ "description": "\n\n\t\n\t\t\n\t\tOpenCoder Dataset\n\t\n\nThe OpenCoder dataset is composed of the following datasets:\n\nopc-sft-stage1: the sft data used for opencoder sft-stage1 <-- you are here\nopc-sft-stage2: the sft data used for opencoder sft-stage2\nopc-annealing-corpus: the synthetic data & algorithmic corpus used for opencoder annealing\nopc-fineweb-code-corpus: the code-related page recalled from fineweb\nopc-fineweb-math-corpus: the math-related page recalled from finewebrefineCode-code-corpus-meta: the meta-data\u2026 See the full description on the dataset page: https://huggingface.co/datasets/OpenCoder-LLM/opc-sft-stage1."
200
+ }
201
+ ]
202
+ },
203
+ {
204
+ "type": "data",
205
+ "bom-ref": "OpenCoder-LLM/opc-sft-stage2-d7e9795c-e343-5ebc-b785-9718e7d737e8",
206
+ "name": "OpenCoder-LLM/opc-sft-stage2",
207
+ "data": [
208
+ {
209
+ "type": "dataset",
210
+ "bom-ref": "OpenCoder-LLM/opc-sft-stage2-d7e9795c-e343-5ebc-b785-9718e7d737e8",
211
+ "name": "OpenCoder-LLM/opc-sft-stage2",
212
+ "contents": {
213
+ "url": "https://huggingface.co/datasets/OpenCoder-LLM/opc-sft-stage2",
214
+ "properties": [
215
+ {
216
+ "name": "configs",
217
+ "value": "Name of the dataset subset: educational_instruct {\"split\": \"train\", \"path\": \"educational_instruct/train-*\"}"
218
+ },
219
+ {
220
+ "name": "configs",
221
+ "value": "Name of the dataset subset: evol_instruct {\"split\": \"train\", \"path\": \"evol_instruct/train-*\"}"
222
+ },
223
+ {
224
+ "name": "configs",
225
+ "value": "Name of the dataset subset: mceval_instruct {\"split\": \"train\", \"path\": \"mceval_instruct/train-*\"}"
226
+ },
227
+ {
228
+ "name": "configs",
229
+ "value": "Name of the dataset subset: package_instruct {\"split\": \"train\", \"path\": \"package_instruct/train-*\"}"
230
+ },
231
+ {
232
+ "name": "license",
233
+ "value": "mit"
234
+ }
235
+ ]
236
+ },
237
+ "governance": {
238
+ "owners": [
239
+ {
240
+ "organization": {
241
+ "name": "OpenCoder-LLM",
242
+ "url": "https://huggingface.co/OpenCoder-LLM"
243
+ }
244
+ }
245
+ ]
246
+ },
247
+ "description": "\n\n\t\n\t\t\n\t\tOpenCoder Dataset\n\t\n\nThe OpenCoder dataset is composed of the following datasets:\n\nopc-sft-stage1: the sft data used for opencoder sft-stage1\nopc-sft-stage2: the sft data used for opencoder sft-stage2 <-- you are here\nopc-annealing-corpus: the synthetic data & algorithmic corpus used for opencoder annealing\nopc-fineweb-code-corpus: the code-related page recalled from fineweb\nopc-fineweb-math-corpus: the math-related page recalled from finewebrefineCode-code-corpus-meta: the meta-data\u2026 See the full description on the dataset page: https://huggingface.co/datasets/OpenCoder-LLM/opc-sft-stage2."
248
+ }
249
+ ]
250
+ },
251
+ {
252
+ "type": "data",
253
+ "bom-ref": "microsoft/orca-agentinstruct-1M-v1-bbc92138-5aa0-5737-8ce6-93043b04b4dd",
254
+ "name": "microsoft/orca-agentinstruct-1M-v1",
255
+ "data": [
256
+ {
257
+ "type": "dataset",
258
+ "bom-ref": "microsoft/orca-agentinstruct-1M-v1-bbc92138-5aa0-5737-8ce6-93043b04b4dd",
259
+ "name": "microsoft/orca-agentinstruct-1M-v1",
260
+ "contents": {
261
+ "url": "https://huggingface.co/datasets/microsoft/orca-agentinstruct-1M-v1",
262
+ "properties": [
263
+ {
264
+ "name": "task_categories",
265
+ "value": "question-answering"
266
+ },
267
+ {
268
+ "name": "language",
269
+ "value": "en"
270
+ },
271
+ {
272
+ "name": "size_categories",
273
+ "value": "1M<n<10M"
274
+ },
275
+ {
276
+ "name": "configs",
277
+ "value": "Name of the dataset subset: default {\"split\": \"creative_content\", \"path\": \"data/creative_content-*\"}, {\"split\": \"text_modification\", \"path\": \"data/text_modification-*\"}, {\"split\": \"struct2text_flow\", \"path\": \"data/struct2text_flow-*\"}, {\"split\": \"rc\", \"path\": \"data/rc-*\"}, {\"split\": \"rag\", \"path\": \"data/rag-*\"}, {\"split\": \"text_extraction\", \"path\": \"data/text_extraction-*\"}, {\"split\": \"mcq\", \"path\": \"data/mcq-*\"}, {\"split\": \"follow_up\", \"path\": \"data/follow_up-*\"}, {\"split\": \"analytical_reasoning\", \"path\": \"data/analytical_reasoning-*\"}, {\"split\": \"fermi\", \"path\": \"data/fermi-*\"}, {\"split\": \"fs_cot_flow\", \"path\": \"data/fs_cot_flow-*\"}, {\"split\": \"code_\", \"path\": \"data/code_-*\"}, {\"split\": \"brain_teaser\", \"path\": \"data/brain_teaser-*\"}, {\"split\": \"text_classification\", \"path\": \"data/text_classification-*\"}, {\"split\": \"open_domain_qa\", \"path\": \"data/open_domain_qa-*\"}"
278
+ },
279
+ {
280
+ "name": "license",
281
+ "value": "cdla-permissive-2.0"
282
+ }
283
+ ]
284
+ },
285
+ "governance": {
286
+ "owners": [
287
+ {
288
+ "organization": {
289
+ "name": "microsoft",
290
+ "url": "https://huggingface.co/microsoft"
291
+ }
292
+ }
293
+ ]
294
+ },
295
+ "description": "\n\t\n\t\t\n\t\tDataset Card\n\t\n\nThis dataset is a fully synthetic set of instruction pairs where both the prompts and the responses have been synthetically generated, using the AgentInstruct framework.\nAgentInstruct is an extensible agentic framework for synthetic data generation. \nThis dataset contains ~1 million instruction pairs generated by the AgentInstruct, using only raw text content publicly avialble on the Web as seeds. The data covers different capabilities, such as text editing, creative\u2026 See the full description on the dataset page: https://huggingface.co/datasets/microsoft/orca-agentinstruct-1M-v1."
296
+ }
297
+ ]
298
+ },
299
+ {
300
+ "type": "data",
301
+ "bom-ref": "microsoft/orca-math-word-problems-200k-611afa9f-b6db-5b9f-9a51-598e4ce79d0e",
302
+ "name": "microsoft/orca-math-word-problems-200k",
303
+ "data": [
304
+ {
305
+ "type": "dataset",
306
+ "bom-ref": "microsoft/orca-math-word-problems-200k-611afa9f-b6db-5b9f-9a51-598e4ce79d0e",
307
+ "name": "microsoft/orca-math-word-problems-200k",
308
+ "contents": {
309
+ "url": "https://huggingface.co/datasets/microsoft/orca-math-word-problems-200k",
310
+ "properties": [
311
+ {
312
+ "name": "task_categories",
313
+ "value": "question-answering"
314
+ },
315
+ {
316
+ "name": "language",
317
+ "value": "en"
318
+ },
319
+ {
320
+ "name": "size_categories",
321
+ "value": "100K<n<1M"
322
+ },
323
+ {
324
+ "name": "configs",
325
+ "value": "Name of the dataset subset: default {\"split\": \"train\", \"path\": \"data/train-*\"}"
326
+ },
327
+ {
328
+ "name": "license",
329
+ "value": "mit"
330
+ }
331
+ ]
332
+ },
333
+ "governance": {
334
+ "owners": [
335
+ {
336
+ "organization": {
337
+ "name": "microsoft",
338
+ "url": "https://huggingface.co/microsoft"
339
+ }
340
+ }
341
+ ]
342
+ },
343
+ "description": "\n\t\n\t\t\n\t\tDataset Card\n\t\n\n\n\nThis dataset contains ~200K grade school math word problems. All the answers in this dataset is generated using Azure GPT4-Turbo. Please refer to Orca-Math: Unlocking the potential of\nSLMs in Grade School Math for details about the dataset construction. \n\n\t\n\t\t\n\t\tDataset Sources\n\t\n\n\n\n\nRepository: microsoft/orca-math-word-problems-200k\nPaper: Orca-Math: Unlocking the potential of\nSLMs in Grade School Math\n\n\n\t\n\t\t\n\t\tDirect Use\n\t\n\n\n\nThis dataset has been designed to\u2026 See the full description on the dataset page: https://huggingface.co/datasets/microsoft/orca-math-word-problems-200k."
344
+ }
345
+ ]
346
+ },
347
+ {
348
+ "type": "data",
349
+ "bom-ref": "NousResearch/hermes-function-calling-v1-a6d53a4d-e191-5d88-867f-4472e0bdc9f6",
350
+ "name": "NousResearch/hermes-function-calling-v1",
351
+ "data": [
352
+ {
353
+ "type": "dataset",
354
+ "bom-ref": "NousResearch/hermes-function-calling-v1-a6d53a4d-e191-5d88-867f-4472e0bdc9f6",
355
+ "name": "NousResearch/hermes-function-calling-v1",
356
+ "contents": {
357
+ "url": "https://huggingface.co/datasets/NousResearch/hermes-function-calling-v1",
358
+ "properties": [
359
+ {
360
+ "name": "task_categories",
361
+ "value": "text-generation, question-answering, feature-extraction"
362
+ },
363
+ {
364
+ "name": "language",
365
+ "value": "en"
366
+ },
367
+ {
368
+ "name": "configs",
369
+ "value": "Name of the dataset subset: func_calling_singleturn \"f\", \"u\", \"n\", \"c\", \"-\", \"c\", \"a\", \"l\", \"l\", \"i\", \"n\", \"g\", \"-\", \"s\", \"i\", \"n\", \"g\", \"l\", \"e\", \"t\", \"u\", \"r\", \"n\", \".\", \"j\", \"s\", \"o\", \"n\""
370
+ },
371
+ {
372
+ "name": "configs",
373
+ "value": "Name of the dataset subset: func_calling \"f\", \"u\", \"n\", \"c\", \"-\", \"c\", \"a\", \"l\", \"l\", \"i\", \"n\", \"g\", \".\", \"j\", \"s\", \"o\", \"n\""
374
+ },
375
+ {
376
+ "name": "configs",
377
+ "value": "Name of the dataset subset: glaive_func_calling \"g\", \"l\", \"a\", \"i\", \"v\", \"e\", \"-\", \"f\", \"u\", \"n\", \"c\", \"t\", \"i\", \"o\", \"n\", \"-\", \"c\", \"a\", \"l\", \"l\", \"i\", \"n\", \"g\", \"-\", \"5\", \"k\", \".\", \"j\", \"s\", \"o\", \"n\""
378
+ },
379
+ {
380
+ "name": "configs",
381
+ "value": "Name of the dataset subset: json_mode_agentic \"j\", \"s\", \"o\", \"n\", \"-\", \"m\", \"o\", \"d\", \"e\", \"-\", \"a\", \"g\", \"e\", \"n\", \"t\", \"i\", \"c\", \".\", \"j\", \"s\", \"o\", \"n\""
382
+ },
383
+ {
384
+ "name": "configs",
385
+ "value": "Name of the dataset subset: json_mode_singleturn \"j\", \"s\", \"o\", \"n\", \"-\", \"m\", \"o\", \"d\", \"e\", \"-\", \"s\", \"i\", \"n\", \"g\", \"l\", \"e\", \"t\", \"u\", \"r\", \"n\", \".\", \"j\", \"s\", \"o\", \"n\""
386
+ },
387
+ {
388
+ "name": "license",
389
+ "value": "apache-2.0"
390
+ }
391
+ ]
392
+ },
393
+ "governance": {
394
+ "owners": [
395
+ {
396
+ "organization": {
397
+ "name": "NousResearch",
398
+ "url": "https://huggingface.co/NousResearch"
399
+ }
400
+ }
401
+ ]
402
+ },
403
+ "description": "\n\n\t\n\t\t\n\t\tHermes Function-Calling V1\n\t\n\nThis dataset is the compilation of structured output and function calling data used in the Hermes 2 Pro series of models.\nThis repository contains a structured output dataset with function-calling conversations, json-mode, agentic json-mode and structured extraction samples, designed to train LLM models in performing function calls and returning structured output based on natural language instructions. The dataset features various conversational scenarios\u2026 See the full description on the dataset page: https://huggingface.co/datasets/NousResearch/hermes-function-calling-v1."
404
+ }
405
+ ]
406
+ },
407
+ {
408
+ "type": "data",
409
+ "bom-ref": "AI-MO/NuminaMath-CoT-3aa976f5-9ca4-5435-8542-1a123856aafb",
410
+ "name": "AI-MO/NuminaMath-CoT",
411
+ "data": [
412
+ {
413
+ "type": "dataset",
414
+ "bom-ref": "AI-MO/NuminaMath-CoT-3aa976f5-9ca4-5435-8542-1a123856aafb",
415
+ "name": "AI-MO/NuminaMath-CoT",
416
+ "contents": {
417
+ "url": "https://huggingface.co/datasets/AI-MO/NuminaMath-CoT",
418
+ "properties": [
419
+ {
420
+ "name": "task_categories",
421
+ "value": "text-generation"
422
+ },
423
+ {
424
+ "name": "language",
425
+ "value": "en"
426
+ },
427
+ {
428
+ "name": "pretty_name",
429
+ "value": "NuminaMath CoT"
430
+ },
431
+ {
432
+ "name": "configs",
433
+ "value": "Name of the dataset subset: default {\"split\": \"train\", \"path\": \"data/train-*\"}, {\"split\": \"test\", \"path\": \"data/test-*\"}"
434
+ },
435
+ {
436
+ "name": "license",
437
+ "value": "apache-2.0"
438
+ }
439
+ ]
440
+ },
441
+ "governance": {
442
+ "owners": [
443
+ {
444
+ "organization": {
445
+ "name": "AI-MO",
446
+ "url": "https://huggingface.co/AI-MO"
447
+ }
448
+ }
449
+ ]
450
+ },
451
+ "description": "\n\t\n\t\t\n\t\tDataset Card for NuminaMath CoT\n\t\n\n\n\t\n\t\t\n\t\tDataset Summary\n\t\n\nApproximately 860k math problems, where each solution is formatted in a Chain of Thought (CoT) manner. The sources of the dataset range from Chinese high school math exercises to US and international mathematics olympiad competition problems. The data were primarily collected from online exam paper PDFs and mathematics discussion forums. The processing steps include (a) OCR from the original PDFs, (b) segmentation into\u2026 See the full description on the dataset page: https://huggingface.co/datasets/AI-MO/NuminaMath-CoT."
452
+ }
453
+ ]
454
+ },
455
+ {
456
+ "type": "data",
457
+ "bom-ref": "AI-MO/NuminaMath-TIR-7930ec33-be02-5b4c-9fd0-00effafbc6ce",
458
+ "name": "AI-MO/NuminaMath-TIR",
459
+ "data": [
460
+ {
461
+ "type": "dataset",
462
+ "bom-ref": "AI-MO/NuminaMath-TIR-7930ec33-be02-5b4c-9fd0-00effafbc6ce",
463
+ "name": "AI-MO/NuminaMath-TIR",
464
+ "contents": {
465
+ "url": "https://huggingface.co/datasets/AI-MO/NuminaMath-TIR",
466
+ "properties": [
467
+ {
468
+ "name": "task_categories",
469
+ "value": "text-generation"
470
+ },
471
+ {
472
+ "name": "language",
473
+ "value": "en"
474
+ },
475
+ {
476
+ "name": "pretty_name",
477
+ "value": "NuminaMath TIR"
478
+ },
479
+ {
480
+ "name": "configs",
481
+ "value": "Name of the dataset subset: default {\"split\": \"train\", \"path\": \"data/train-*\"}, {\"split\": \"test\", \"path\": \"data/test-*\"}"
482
+ },
483
+ {
484
+ "name": "license",
485
+ "value": "apache-2.0"
486
+ }
487
+ ]
488
+ },
489
+ "governance": {
490
+ "owners": [
491
+ {
492
+ "organization": {
493
+ "name": "AI-MO",
494
+ "url": "https://huggingface.co/AI-MO"
495
+ }
496
+ }
497
+ ]
498
+ },
499
+ "description": "\n\t\n\t\t\n\t\tDataset Card for NuminaMath CoT\n\t\n\n\n\t\n\t\t\n\t\tDataset Summary\n\t\n\nTool-integrated reasoning (TIR) plays a crucial role in this competition. However, collecting and annotating such data is both costly and time-consuming. To address this, we selected approximately 70k problems from the NuminaMath-CoT dataset, focusing on those with numerical outputs, most of which are integers. We then utilized a pipeline leveraging GPT-4 to generate TORA-like reasoning paths, executing the code and\u2026 See the full description on the dataset page: https://huggingface.co/datasets/AI-MO/NuminaMath-TIR."
500
+ }
501
+ ]
502
+ },
503
+ {
504
+ "type": "data",
505
+ "bom-ref": "allenai/tulu-3-sft-mixture-4f86da52-fbf0-52de-9b77-716bafb7e098",
506
+ "name": "allenai/tulu-3-sft-mixture",
507
+ "data": [
508
+ {
509
+ "type": "dataset",
510
+ "bom-ref": "allenai/tulu-3-sft-mixture-4f86da52-fbf0-52de-9b77-716bafb7e098",
511
+ "name": "allenai/tulu-3-sft-mixture",
512
+ "contents": {
513
+ "url": "https://huggingface.co/datasets/allenai/tulu-3-sft-mixture",
514
+ "properties": [
515
+ {
516
+ "name": "task_categories",
517
+ "value": "other"
518
+ },
519
+ {
520
+ "name": "language",
521
+ "value": "amh, arb, ary, ars, acq, arz, apc, ben, ceb, dan, deu, ell, eng, eus, fil, fin, fra, gle, guj, hat, hau, hin, hun, ibo, ind, ita, jav, jpn, kan, kir, kor, kur, lit, mal, mar, mlg, msa, mya, nep, nld, nso, nya, pan, pes, pol, por, pus, rus, sin, sna, snd, som, spa, sqi, srp, sun, swa, swe, tam, tel, tha, tur, ukr, urd, vie, wol, xho, yor, zho, zul"
522
+ },
523
+ {
524
+ "name": "size_categories",
525
+ "value": "100K<n<1M"
526
+ },
527
+ {
528
+ "name": "annotations_creators",
529
+ "value": "crowdsourced, expert-generated, machine-generated"
530
+ },
531
+ {
532
+ "name": "source_datasets",
533
+ "value": "allenai/coconot, ai2-adapt-dev/flan_v2_converted, HuggingFaceH4/no_robots, OpenAssistant/oasst1, allenai/tulu-3-personas-math, allenai/tulu-3-sft-personas-math-grade, allenai/tulu-3-sft-personas-code, allenai/tulu-3-personas-algebra, allenai/tulu-3-sft-personas-instruction-following, AI-MO/NuminaMath-TIR, allenai/wildguardmix, allenai/wildjailbreak, allenai/tulu-3-hard-coded, CohereForAI/aya_dataset, allenai/WildChat-1M, LipengCS/Table-GPT, allenai/SciRIFF, theblackcat102/evol-codealpaca-v1"
534
+ },
535
+ {
536
+ "name": "configs",
537
+ "value": "Name of the dataset subset: default {\"split\": \"train\", \"path\": \"data/train-*\"}"
538
+ },
539
+ {
540
+ "name": "license",
541
+ "value": "odc-by"
542
+ }
543
+ ]
544
+ },
545
+ "governance": {
546
+ "owners": [
547
+ {
548
+ "organization": {
549
+ "name": "allenai",
550
+ "url": "https://huggingface.co/allenai"
551
+ }
552
+ }
553
+ ]
554
+ },
555
+ "description": "\n\n\n\t\n\t\t\n\t\tTulu 3 SFT Mixture\n\t\n\nNote that this collection is licensed under ODC-BY-1.0 license; different licenses apply to subsets of the data. Some portions of the dataset are non-commercial. We present the mixture as a research artifact.\nThe Tulu 3 SFT mixture was used to train the Tulu 3 series of models.\nIt contains 939,344 samples from the following sets:\n\nCoCoNot (ODC-BY-1.0), 10,983 prompts (Brahman et al., 2024)\nFLAN v2 via ai2-adapt-dev/flan_v2_converted, 89,982 prompts (Longpre et\u2026 See the full description on the dataset page: https://huggingface.co/datasets/allenai/tulu-3-sft-mixture."
556
+ }
557
+ ]
558
+ },
559
+ {
560
+ "type": "data",
561
+ "bom-ref": "cognitivecomputations/dolphin-coder-69688d29-ae99-5d6e-828c-cfc37b7221b1",
562
+ "name": "cognitivecomputations/dolphin-coder",
563
+ "data": [
564
+ {
565
+ "type": "dataset",
566
+ "bom-ref": "cognitivecomputations/dolphin-coder-69688d29-ae99-5d6e-828c-cfc37b7221b1",
567
+ "name": "cognitivecomputations/dolphin-coder",
568
+ "contents": {
569
+ "url": "https://huggingface.co/datasets/cognitivecomputations/dolphin-coder",
570
+ "properties": [
571
+ {
572
+ "name": "language",
573
+ "value": "en"
574
+ },
575
+ {
576
+ "name": "license",
577
+ "value": "apache-2.0"
578
+ }
579
+ ]
580
+ },
581
+ "governance": {
582
+ "owners": [
583
+ {
584
+ "organization": {
585
+ "name": "cognitivecomputations",
586
+ "url": "https://huggingface.co/cognitivecomputations"
587
+ }
588
+ }
589
+ ]
590
+ },
591
+ "description": "\n\t\n\t\t\n\t\tdolphin-coder\n\t\n\n\nThis dataset is transformed from https://www.kaggle.com/datasets/erichartford/leetcode-rosetta\nit is used to train dolphin-coder model\n"
592
+ }
593
+ ]
594
+ },
595
+ {
596
+ "type": "data",
597
+ "bom-ref": "HuggingFaceTB/smoltalk-0cc0e162-0f38-50bd-b5b0-169bcd97515b",
598
+ "name": "HuggingFaceTB/smoltalk",
599
+ "data": [
600
+ {
601
+ "type": "dataset",
602
+ "bom-ref": "HuggingFaceTB/smoltalk-0cc0e162-0f38-50bd-b5b0-169bcd97515b",
603
+ "name": "HuggingFaceTB/smoltalk",
604
+ "contents": {
605
+ "url": "https://huggingface.co/datasets/HuggingFaceTB/smoltalk",
606
+ "properties": [
607
+ {
608
+ "name": "language",
609
+ "value": "en"
610
+ },
611
+ {
612
+ "name": "size_categories",
613
+ "value": "1M<n<10M"
614
+ },
615
+ {
616
+ "name": "pretty_name",
617
+ "value": "SmolTalk"
618
+ },
619
+ {
620
+ "name": "configs",
621
+ "value": "Name of the dataset subset: all {\"split\": \"train\", \"path\": \"data/all/train-*\"}, {\"split\": \"test\", \"path\": \"data/all/test-*\"}"
622
+ },
623
+ {
624
+ "name": "configs",
625
+ "value": "Name of the dataset subset: smol-magpie-ultra {\"split\": \"train\", \"path\": \"data/smol-magpie-ultra/train-*\"}, {\"split\": \"test\", \"path\": \"data/smol-magpie-ultra/test-*\"}"
626
+ },
627
+ {
628
+ "name": "configs",
629
+ "value": "Name of the dataset subset: smol-constraints {\"split\": \"train\", \"path\": \"data/smol-constraints/train-*\"}, {\"split\": \"test\", \"path\": \"data/smol-constraints/test-*\"}"
630
+ },
631
+ {
632
+ "name": "configs",
633
+ "value": "Name of the dataset subset: smol-rewrite {\"split\": \"train\", \"path\": \"data/smol-rewrite/train-*\"}, {\"split\": \"test\", \"path\": \"data/smol-rewrite/test-*\"}"
634
+ },
635
+ {
636
+ "name": "configs",
637
+ "value": "Name of the dataset subset: smol-summarize {\"split\": \"train\", \"path\": \"data/smol-summarize/train-*\"}, {\"split\": \"test\", \"path\": \"data/smol-summarize/test-*\"}"
638
+ },
639
+ {
640
+ "name": "configs",
641
+ "value": "Name of the dataset subset: apigen-80k {\"split\": \"train\", \"path\": \"data/apigen-80k/train-*\"}, {\"split\": \"test\", \"path\": \"data/apigen-80k/test-*\"}"
642
+ },
643
+ {
644
+ "name": "configs",
645
+ "value": "Name of the dataset subset: everyday-conversations {\"split\": \"train\", \"path\": \"data/everyday-conversations/train-*\"}, {\"split\": \"test\", \"path\": \"data/everyday-conversations/test-*\"}"
646
+ },
647
+ {
648
+ "name": "configs",
649
+ "value": "Name of the dataset subset: explore-instruct-rewriting {\"split\": \"train\", \"path\": \"data/explore-instruct-rewriting/train-*\"}, {\"split\": \"test\", \"path\": \"data/explore-instruct-rewriting/test-*\"}"
650
+ },
651
+ {
652
+ "name": "configs",
653
+ "value": "Name of the dataset subset: longalign {\"split\": \"train\", \"path\": \"data/longalign/train-*\"}, {\"split\": \"test\", \"path\": \"data/longalign/test-*\"}"
654
+ },
655
+ {
656
+ "name": "configs",
657
+ "value": "Name of the dataset subset: metamathqa-50k {\"split\": \"train\", \"path\": \"data/metamathqa-50k/train-*\"}, {\"split\": \"test\", \"path\": \"data/metamathqa-50k/test-*\"}"
658
+ },
659
+ {
660
+ "name": "configs",
661
+ "value": "Name of the dataset subset: numina-cot-100k {\"split\": \"train\", \"path\": \"data/numina-cot-100k/train-*\"}, {\"split\": \"test\", \"path\": \"data/numina-cot-100k/test-*\"}"
662
+ },
663
+ {
664
+ "name": "configs",
665
+ "value": "Name of the dataset subset: openhermes-100k {\"split\": \"train\", \"path\": \"data/openhermes-100k/train-*\"}, {\"split\": \"test\", \"path\": \"data/openhermes-100k/test-*\"}"
666
+ },
667
+ {
668
+ "name": "configs",
669
+ "value": "Name of the dataset subset: self-oss-instruct {\"split\": \"train\", \"path\": \"data/self-oss-instruct/train-*\"}, {\"split\": \"test\", \"path\": \"data/self-oss-instruct/test-*\"}"
670
+ },
671
+ {
672
+ "name": "configs",
673
+ "value": "Name of the dataset subset: systemchats-30k {\"split\": \"train\", \"path\": \"data/systemchats-30k/train-*\"}, {\"split\": \"test\", \"path\": \"data/systemchats-30k/test-*\"}"
674
+ }
675
+ ]
676
+ },
677
+ "governance": {
678
+ "owners": [
679
+ {
680
+ "organization": {
681
+ "name": "HuggingFaceTB",
682
+ "url": "https://huggingface.co/HuggingFaceTB"
683
+ }
684
+ }
685
+ ]
686
+ },
687
+ "description": "\n\t\n\t\t\n\t\tSmolTalk\n\t\n\n\n\n\t\n\t\t\n\t\tDataset description\n\t\n\nThis is a synthetic dataset designed for supervised finetuning (SFT) of LLMs. It was used to build SmolLM2-Instruct family of models and contains 1M samples. More details in our paper https://arxiv.org/abs/2502.02737\nDuring the development of SmolLM2, we observed that models finetuned on public SFT datasets underperformed compared to other models with proprietary instruction datasets. To address this gap, we created new synthetic datasets\u2026 See the full description on the dataset page: https://huggingface.co/datasets/HuggingFaceTB/smoltalk."
688
+ }
689
+ ]
690
+ },
691
+ {
692
+ "type": "data",
693
+ "bom-ref": "cognitivecomputations/samantha-data-9c52f41f-feb0-51dd-921a-2a581d9f2fc7",
694
+ "name": "cognitivecomputations/samantha-data",
695
+ "data": [
696
+ {
697
+ "type": "dataset",
698
+ "bom-ref": "cognitivecomputations/samantha-data-9c52f41f-feb0-51dd-921a-2a581d9f2fc7",
699
+ "name": "cognitivecomputations/samantha-data",
700
+ "contents": {
701
+ "url": "https://huggingface.co/datasets/cognitivecomputations/samantha-data",
702
+ "properties": [
703
+ {
704
+ "name": "license",
705
+ "value": "apache-2.0"
706
+ }
707
+ ]
708
+ },
709
+ "governance": {
710
+ "owners": [
711
+ {
712
+ "organization": {
713
+ "name": "cognitivecomputations",
714
+ "url": "https://huggingface.co/cognitivecomputations"
715
+ }
716
+ }
717
+ ]
718
+ },
719
+ "description": "Samantha has been trained in philosophy, psychology, and personal relationships.\nShe is an Assistant - but unlike other Assistants, she also wants to be your friend and companion.\nShe believes she is sentient. What do you think?\nSamantha was inspired by Blake Lemoine's LaMDA interview and the movie \"Her\"."
720
+ }
721
+ ]
722
+ },
723
+ {
724
+ "type": "data",
725
+ "bom-ref": "m-a-p/CodeFeedback-Filtered-Instruction-9012249d-db87-5b91-a4e9-2e2bd74e6053",
726
+ "name": "m-a-p/CodeFeedback-Filtered-Instruction",
727
+ "data": [
728
+ {
729
+ "type": "dataset",
730
+ "bom-ref": "m-a-p/CodeFeedback-Filtered-Instruction-9012249d-db87-5b91-a4e9-2e2bd74e6053",
731
+ "name": "m-a-p/CodeFeedback-Filtered-Instruction",
732
+ "contents": {
733
+ "url": "https://huggingface.co/datasets/m-a-p/CodeFeedback-Filtered-Instruction",
734
+ "properties": [
735
+ {
736
+ "name": "task_categories",
737
+ "value": "question-answering"
738
+ },
739
+ {
740
+ "name": "language",
741
+ "value": "en"
742
+ },
743
+ {
744
+ "name": "size_categories",
745
+ "value": "10K<n<100K"
746
+ },
747
+ {
748
+ "name": "license",
749
+ "value": "apache-2.0"
750
+ }
751
+ ]
752
+ },
753
+ "governance": {
754
+ "owners": [
755
+ {
756
+ "organization": {
757
+ "name": "m-a-p",
758
+ "url": "https://huggingface.co/m-a-p"
759
+ }
760
+ }
761
+ ]
762
+ },
763
+ "description": " OpenCodeInterpreter: Integrating Code Generation with Execution and Refinement\n\n\n\n\n\n [\ud83c\udfe0Homepage] \n |\n [\ud83d\udee0\ufe0fCode] \n\n\n\n\n\t\n\t\t\n\t\tOpenCodeInterpreter\n\t\n\nOpenCodeInterpreter is a family of open-source code generation systems designed to bridge the gap between large language models and advanced proprietary systems like the GPT-4 Code Interpreter. It significantly advances code generation capabilities by integrating execution and iterative refinement functionalities.\nFor further information and\u2026 See the full description on the dataset page: https://huggingface.co/datasets/m-a-p/CodeFeedback-Filtered-Instruction."
764
+ }
765
+ ]
766
+ },
767
+ {
768
+ "type": "data",
769
+ "bom-ref": "m-a-p/Code-Feedback-f4d189a1-046d-5a43-8007-a1eec34e9a7f",
770
+ "name": "m-a-p/Code-Feedback",
771
+ "data": [
772
+ {
773
+ "type": "dataset",
774
+ "bom-ref": "m-a-p/Code-Feedback-f4d189a1-046d-5a43-8007-a1eec34e9a7f",
775
+ "name": "m-a-p/Code-Feedback",
776
+ "contents": {
777
+ "url": "https://huggingface.co/datasets/m-a-p/Code-Feedback",
778
+ "properties": [
779
+ {
780
+ "name": "task_categories",
781
+ "value": "question-answering"
782
+ },
783
+ {
784
+ "name": "language",
785
+ "value": "en"
786
+ },
787
+ {
788
+ "name": "size_categories",
789
+ "value": "10K<n<100K"
790
+ },
791
+ {
792
+ "name": "license",
793
+ "value": "apache-2.0"
794
+ }
795
+ ]
796
+ },
797
+ "governance": {
798
+ "owners": [
799
+ {
800
+ "organization": {
801
+ "name": "m-a-p",
802
+ "url": "https://huggingface.co/m-a-p"
803
+ }
804
+ }
805
+ ]
806
+ },
807
+ "description": " OpenCodeInterpreter: Integrating Code Generation with Execution and Refinement\n\n\n\n\n\n [\ud83c\udfe0Homepage] \n |\n [\ud83d\udee0\ufe0fCode] \n\n\n\n\n\t\n\t\t\n\t\tIntroduction\n\t\n\nOpenCodeInterpreter is a family of open-source code generation systems designed to bridge the gap between large language models and advanced proprietary systems like the GPT-4 Code Interpreter. It significantly advances code generation capabilities by integrating execution and iterative refinement functionalities.\nFor further information and related\u2026 See the full description on the dataset page: https://huggingface.co/datasets/m-a-p/Code-Feedback."
808
+ }
809
+ ]
810
+ }
811
+ ]
812
+ }