cognitivecomputations_Dolphin3.0-R1-Mistral-24B.json ADDED
@@ -0,0 +1,812 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bomFormat": "CycloneDX",
3
+ "specVersion": "1.6",
4
+ "serialNumber": "urn:uuid:e6ef2cd0-b2c1-4b2e-ba7c-adb01c441786",
5
+ "version": 1,
6
+ "metadata": {
7
+ "timestamp": "2025-06-05T09:34:41.455909+00:00",
8
+ "component": {
9
+ "type": "machine-learning-model",
10
+ "bom-ref": "cognitivecomputations/Dolphin3.0-R1-Mistral-24B-6784ad22-545f-555e-aac7-abd08a7c4fc0",
11
+ "name": "cognitivecomputations/Dolphin3.0-R1-Mistral-24B",
12
+ "externalReferences": [
13
+ {
14
+ "url": "https://huggingface.co/cognitivecomputations/Dolphin3.0-R1-Mistral-24B",
15
+ "type": "documentation"
16
+ }
17
+ ],
18
+ "modelCard": {
19
+ "modelParameters": {
20
+ "task": "text-generation",
21
+ "architectureFamily": "mistral",
22
+ "modelArchitecture": "MistralForCausalLM",
23
+ "datasets": [
24
+ {
25
+ "ref": "cognitivecomputations/dolphin-r1-e86781ae-297a-5171-8a0d-61460dd40738"
26
+ },
27
+ {
28
+ "ref": "OpenCoder-LLM/opc-sft-stage1-f0572249-17d5-5847-8e5a-72583934eca0"
29
+ },
30
+ {
31
+ "ref": "OpenCoder-LLM/opc-sft-stage2-d7e9795c-e343-5ebc-b785-9718e7d737e8"
32
+ },
33
+ {
34
+ "ref": "microsoft/orca-agentinstruct-1M-v1-bbc92138-5aa0-5737-8ce6-93043b04b4dd"
35
+ },
36
+ {
37
+ "ref": "microsoft/orca-math-word-problems-200k-611afa9f-b6db-5b9f-9a51-598e4ce79d0e"
38
+ },
39
+ {
40
+ "ref": "NousResearch/hermes-function-calling-v1-a6d53a4d-e191-5d88-867f-4472e0bdc9f6"
41
+ },
42
+ {
43
+ "ref": "AI-MO/NuminaMath-CoT-3aa976f5-9ca4-5435-8542-1a123856aafb"
44
+ },
45
+ {
46
+ "ref": "AI-MO/NuminaMath-TIR-7930ec33-be02-5b4c-9fd0-00effafbc6ce"
47
+ },
48
+ {
49
+ "ref": "allenai/tulu-3-sft-mixture-4f86da52-fbf0-52de-9b77-716bafb7e098"
50
+ },
51
+ {
52
+ "ref": "cognitivecomputations/dolphin-coder-69688d29-ae99-5d6e-828c-cfc37b7221b1"
53
+ },
54
+ {
55
+ "ref": "HuggingFaceTB/smoltalk-0cc0e162-0f38-50bd-b5b0-169bcd97515b"
56
+ },
57
+ {
58
+ "ref": "cognitivecomputations/samantha-data-9c52f41f-feb0-51dd-921a-2a581d9f2fc7"
59
+ },
60
+ {
61
+ "ref": "m-a-p/CodeFeedback-Filtered-Instruction-9012249d-db87-5b91-a4e9-2e2bd74e6053"
62
+ },
63
+ {
64
+ "ref": "m-a-p/Code-Feedback-f4d189a1-046d-5a43-8007-a1eec34e9a7f"
65
+ }
66
+ ]
67
+ },
68
+ "properties": [
69
+ {
70
+ "name": "library_name",
71
+ "value": "transformers"
72
+ },
73
+ {
74
+ "name": "base_model",
75
+ "value": "mistralai/Mistral-Small-24B-Base-2501"
76
+ }
77
+ ]
78
+ },
79
+ "authors": [
80
+ {
81
+ "name": "cognitivecomputations"
82
+ }
83
+ ],
84
+ "tags": [
85
+ "transformers",
86
+ "safetensors",
87
+ "mistral",
88
+ "text-generation",
89
+ "conversational",
90
+ "en",
91
+ "dataset:cognitivecomputations/dolphin-r1",
92
+ "dataset:OpenCoder-LLM/opc-sft-stage1",
93
+ "dataset:OpenCoder-LLM/opc-sft-stage2",
94
+ "dataset:microsoft/orca-agentinstruct-1M-v1",
95
+ "dataset:microsoft/orca-math-word-problems-200k",
96
+ "dataset:NousResearch/hermes-function-calling-v1",
97
+ "dataset:AI-MO/NuminaMath-CoT",
98
+ "dataset:AI-MO/NuminaMath-TIR",
99
+ "dataset:allenai/tulu-3-sft-mixture",
100
+ "dataset:cognitivecomputations/dolphin-coder",
101
+ "dataset:HuggingFaceTB/smoltalk",
102
+ "dataset:cognitivecomputations/samantha-data",
103
+ "dataset:m-a-p/CodeFeedback-Filtered-Instruction",
104
+ "dataset:m-a-p/Code-Feedback",
105
+ "base_model:mistralai/Mistral-Small-24B-Base-2501",
106
+ "base_model:finetune:mistralai/Mistral-Small-24B-Base-2501",
107
+ "autotrain_compatible",
108
+ "text-generation-inference",
109
+ "endpoints_compatible",
110
+ "region:us"
111
+ ]
112
+ }
113
+ },
114
+ "components": [
115
+ {
116
+ "type": "data",
117
+ "bom-ref": "cognitivecomputations/dolphin-r1-e86781ae-297a-5171-8a0d-61460dd40738",
118
+ "name": "cognitivecomputations/dolphin-r1",
119
+ "data": [
120
+ {
121
+ "type": "dataset",
122
+ "bom-ref": "cognitivecomputations/dolphin-r1-e86781ae-297a-5171-8a0d-61460dd40738",
123
+ "name": "cognitivecomputations/dolphin-r1",
124
+ "contents": {
125
+ "url": "https://huggingface.co/datasets/cognitivecomputations/dolphin-r1",
126
+ "properties": [
127
+ {
128
+ "name": "configs",
129
+ "value": "Name of the dataset subset: nonreasoning {\"split\": \"train\", \"path\": \"dolphin-r1-nonreasoning.jsonl\"}"
130
+ },
131
+ {
132
+ "name": "configs",
133
+ "value": "Name of the dataset subset: reasoning-deepseek {\"split\": \"train\", \"path\": \"dolphin-r1-reasoning-deepseek.jsonl\"}"
134
+ },
135
+ {
136
+ "name": "configs",
137
+ "value": "Name of the dataset subset: reasoning-flash {\"split\": \"train\", \"path\": \"dolphin-r1-reasoning-flash.jsonl\"}"
138
+ },
139
+ {
140
+ "name": "license",
141
+ "value": "apache-2.0"
142
+ }
143
+ ]
144
+ },
145
+ "governance": {
146
+ "owners": [
147
+ {
148
+ "organization": {
149
+ "name": "cognitivecomputations",
150
+ "url": "https://huggingface.co/cognitivecomputations"
151
+ }
152
+ }
153
+ ]
154
+ },
155
+ "description": "\n\t\n\t\t\n\t\tDolphin R1 \ud83d\udc2c\n\t\n\nAn Apache-2.0 dataset curated by Eric Hartford and Cognitive Computations\n\nDiscord: https://discord.gg/cognitivecomputations\n\n\n\n\t\n\t\t\n\t\n\t\n\t\tSponsors\n\t\n\nOur appreciation for the generous sponsors of Dolphin R1 - Without whom this dataset could not exist.\n\nDria https://x.com/driaforall - Inference Sponsor (DeepSeek)\nChutes https://x.com/rayon_labs - Inference Sponsor (Flash)\nCrusoe Cloud - Compute Sponsor\nAndreessen Horowitz - provided the grant that originally launched\u2026 See the full description on the dataset page: https://huggingface.co/datasets/cognitivecomputations/dolphin-r1."
156
+ }
157
+ ]
158
+ },
159
+ {
160
+ "type": "data",
161
+ "bom-ref": "OpenCoder-LLM/opc-sft-stage1-f0572249-17d5-5847-8e5a-72583934eca0",
162
+ "name": "OpenCoder-LLM/opc-sft-stage1",
163
+ "data": [
164
+ {
165
+ "type": "dataset",
166
+ "bom-ref": "OpenCoder-LLM/opc-sft-stage1-f0572249-17d5-5847-8e5a-72583934eca0",
167
+ "name": "OpenCoder-LLM/opc-sft-stage1",
168
+ "contents": {
169
+ "url": "https://huggingface.co/datasets/OpenCoder-LLM/opc-sft-stage1",
170
+ "properties": [
171
+ {
172
+ "name": "configs",
173
+ "value": "Name of the dataset subset: filtered_infinity_instruct {\"split\": \"train\", \"path\": \"data/filtered_infinity_instruct-*\"}"
174
+ },
175
+ {
176
+ "name": "configs",
177
+ "value": "Name of the dataset subset: largescale_diverse_instruct {\"split\": \"train\", \"path\": \"data/largescale_diverse_instruct-*\"}"
178
+ },
179
+ {
180
+ "name": "configs",
181
+ "value": "Name of the dataset subset: realuser_instruct {\"split\": \"train\", \"path\": \"data/realuser_instruct-*\"}"
182
+ },
183
+ {
184
+ "name": "license",
185
+ "value": "mit"
186
+ }
187
+ ]
188
+ },
189
+ "governance": {
190
+ "owners": [
191
+ {
192
+ "organization": {
193
+ "name": "OpenCoder-LLM",
194
+ "url": "https://huggingface.co/OpenCoder-LLM"
195
+ }
196
+ }
197
+ ]
198
+ },
199
+ "description": "\n\n\t\n\t\t\n\t\tOpenCoder Dataset\n\t\n\nThe OpenCoder dataset is composed of the following datasets:\n\nopc-sft-stage1: the sft data used for opencoder sft-stage1 <-- you are here\nopc-sft-stage2: the sft data used for opencoder sft-stage2\nopc-annealing-corpus: the synthetic data & algorithmic corpus used for opencoder annealing\nopc-fineweb-code-corpus: the code-related page recalled from fineweb\nopc-fineweb-math-corpus: the math-related page recalled from finewebrefineCode-code-corpus-meta: the meta-data\u2026 See the full description on the dataset page: https://huggingface.co/datasets/OpenCoder-LLM/opc-sft-stage1."
200
+ }
201
+ ]
202
+ },
203
+ {
204
+ "type": "data",
205
+ "bom-ref": "OpenCoder-LLM/opc-sft-stage2-d7e9795c-e343-5ebc-b785-9718e7d737e8",
206
+ "name": "OpenCoder-LLM/opc-sft-stage2",
207
+ "data": [
208
+ {
209
+ "type": "dataset",
210
+ "bom-ref": "OpenCoder-LLM/opc-sft-stage2-d7e9795c-e343-5ebc-b785-9718e7d737e8",
211
+ "name": "OpenCoder-LLM/opc-sft-stage2",
212
+ "contents": {
213
+ "url": "https://huggingface.co/datasets/OpenCoder-LLM/opc-sft-stage2",
214
+ "properties": [
215
+ {
216
+ "name": "configs",
217
+ "value": "Name of the dataset subset: educational_instruct {\"split\": \"train\", \"path\": \"educational_instruct/train-*\"}"
218
+ },
219
+ {
220
+ "name": "configs",
221
+ "value": "Name of the dataset subset: evol_instruct {\"split\": \"train\", \"path\": \"evol_instruct/train-*\"}"
222
+ },
223
+ {
224
+ "name": "configs",
225
+ "value": "Name of the dataset subset: mceval_instruct {\"split\": \"train\", \"path\": \"mceval_instruct/train-*\"}"
226
+ },
227
+ {
228
+ "name": "configs",
229
+ "value": "Name of the dataset subset: package_instruct {\"split\": \"train\", \"path\": \"package_instruct/train-*\"}"
230
+ },
231
+ {
232
+ "name": "license",
233
+ "value": "mit"
234
+ }
235
+ ]
236
+ },
237
+ "governance": {
238
+ "owners": [
239
+ {
240
+ "organization": {
241
+ "name": "OpenCoder-LLM",
242
+ "url": "https://huggingface.co/OpenCoder-LLM"
243
+ }
244
+ }
245
+ ]
246
+ },
247
+ "description": "\n\n\t\n\t\t\n\t\tOpenCoder Dataset\n\t\n\nThe OpenCoder dataset is composed of the following datasets:\n\nopc-sft-stage1: the sft data used for opencoder sft-stage1\nopc-sft-stage2: the sft data used for opencoder sft-stage2 <-- you are here\nopc-annealing-corpus: the synthetic data & algorithmic corpus used for opencoder annealing\nopc-fineweb-code-corpus: the code-related page recalled from fineweb\nopc-fineweb-math-corpus: the math-related page recalled from finewebrefineCode-code-corpus-meta: the meta-data\u2026 See the full description on the dataset page: https://huggingface.co/datasets/OpenCoder-LLM/opc-sft-stage2."
248
+ }
249
+ ]
250
+ },
251
+ {
252
+ "type": "data",
253
+ "bom-ref": "microsoft/orca-agentinstruct-1M-v1-bbc92138-5aa0-5737-8ce6-93043b04b4dd",
254
+ "name": "microsoft/orca-agentinstruct-1M-v1",
255
+ "data": [
256
+ {
257
+ "type": "dataset",
258
+ "bom-ref": "microsoft/orca-agentinstruct-1M-v1-bbc92138-5aa0-5737-8ce6-93043b04b4dd",
259
+ "name": "microsoft/orca-agentinstruct-1M-v1",
260
+ "contents": {
261
+ "url": "https://huggingface.co/datasets/microsoft/orca-agentinstruct-1M-v1",
262
+ "properties": [
263
+ {
264
+ "name": "task_categories",
265
+ "value": "question-answering"
266
+ },
267
+ {
268
+ "name": "language",
269
+ "value": "en"
270
+ },
271
+ {
272
+ "name": "size_categories",
273
+ "value": "1M<n<10M"
274
+ },
275
+ {
276
+ "name": "configs",
277
+ "value": "Name of the dataset subset: default {\"split\": \"creative_content\", \"path\": \"data/creative_content-*\"}, {\"split\": \"text_modification\", \"path\": \"data/text_modification-*\"}, {\"split\": \"struct2text_flow\", \"path\": \"data/struct2text_flow-*\"}, {\"split\": \"rc\", \"path\": \"data/rc-*\"}, {\"split\": \"rag\", \"path\": \"data/rag-*\"}, {\"split\": \"text_extraction\", \"path\": \"data/text_extraction-*\"}, {\"split\": \"mcq\", \"path\": \"data/mcq-*\"}, {\"split\": \"follow_up\", \"path\": \"data/follow_up-*\"}, {\"split\": \"analytical_reasoning\", \"path\": \"data/analytical_reasoning-*\"}, {\"split\": \"fermi\", \"path\": \"data/fermi-*\"}, {\"split\": \"fs_cot_flow\", \"path\": \"data/fs_cot_flow-*\"}, {\"split\": \"code_\", \"path\": \"data/code_-*\"}, {\"split\": \"brain_teaser\", \"path\": \"data/brain_teaser-*\"}, {\"split\": \"text_classification\", \"path\": \"data/text_classification-*\"}, {\"split\": \"open_domain_qa\", \"path\": \"data/open_domain_qa-*\"}"
278
+ },
279
+ {
280
+ "name": "license",
281
+ "value": "cdla-permissive-2.0"
282
+ }
283
+ ]
284
+ },
285
+ "governance": {
286
+ "owners": [
287
+ {
288
+ "organization": {
289
+ "name": "microsoft",
290
+ "url": "https://huggingface.co/microsoft"
291
+ }
292
+ }
293
+ ]
294
+ },
295
+ "description": "\n\t\n\t\t\n\t\tDataset Card\n\t\n\nThis dataset is a fully synthetic set of instruction pairs where both the prompts and the responses have been synthetically generated, using the AgentInstruct framework.\nAgentInstruct is an extensible agentic framework for synthetic data generation. \nThis dataset contains ~1 million instruction pairs generated by the AgentInstruct, using only raw text content publicly avialble on the Web as seeds. The data covers different capabilities, such as text editing, creative\u2026 See the full description on the dataset page: https://huggingface.co/datasets/microsoft/orca-agentinstruct-1M-v1."
296
+ }
297
+ ]
298
+ },
299
+ {
300
+ "type": "data",
301
+ "bom-ref": "microsoft/orca-math-word-problems-200k-611afa9f-b6db-5b9f-9a51-598e4ce79d0e",
302
+ "name": "microsoft/orca-math-word-problems-200k",
303
+ "data": [
304
+ {
305
+ "type": "dataset",
306
+ "bom-ref": "microsoft/orca-math-word-problems-200k-611afa9f-b6db-5b9f-9a51-598e4ce79d0e",
307
+ "name": "microsoft/orca-math-word-problems-200k",
308
+ "contents": {
309
+ "url": "https://huggingface.co/datasets/microsoft/orca-math-word-problems-200k",
310
+ "properties": [
311
+ {
312
+ "name": "task_categories",
313
+ "value": "question-answering"
314
+ },
315
+ {
316
+ "name": "language",
317
+ "value": "en"
318
+ },
319
+ {
320
+ "name": "size_categories",
321
+ "value": "100K<n<1M"
322
+ },
323
+ {
324
+ "name": "configs",
325
+ "value": "Name of the dataset subset: default {\"split\": \"train\", \"path\": \"data/train-*\"}"
326
+ },
327
+ {
328
+ "name": "license",
329
+ "value": "mit"
330
+ }
331
+ ]
332
+ },
333
+ "governance": {
334
+ "owners": [
335
+ {
336
+ "organization": {
337
+ "name": "microsoft",
338
+ "url": "https://huggingface.co/microsoft"
339
+ }
340
+ }
341
+ ]
342
+ },
343
+ "description": "\n\t\n\t\t\n\t\tDataset Card\n\t\n\n\n\nThis dataset contains ~200K grade school math word problems. All the answers in this dataset is generated using Azure GPT4-Turbo. Please refer to Orca-Math: Unlocking the potential of\nSLMs in Grade School Math for details about the dataset construction. \n\n\t\n\t\t\n\t\tDataset Sources\n\t\n\n\n\n\nRepository: microsoft/orca-math-word-problems-200k\nPaper: Orca-Math: Unlocking the potential of\nSLMs in Grade School Math\n\n\n\t\n\t\t\n\t\tDirect Use\n\t\n\n\n\nThis dataset has been designed to\u2026 See the full description on the dataset page: https://huggingface.co/datasets/microsoft/orca-math-word-problems-200k."
344
+ }
345
+ ]
346
+ },
347
+ {
348
+ "type": "data",
349
+ "bom-ref": "NousResearch/hermes-function-calling-v1-a6d53a4d-e191-5d88-867f-4472e0bdc9f6",
350
+ "name": "NousResearch/hermes-function-calling-v1",
351
+ "data": [
352
+ {
353
+ "type": "dataset",
354
+ "bom-ref": "NousResearch/hermes-function-calling-v1-a6d53a4d-e191-5d88-867f-4472e0bdc9f6",
355
+ "name": "NousResearch/hermes-function-calling-v1",
356
+ "contents": {
357
+ "url": "https://huggingface.co/datasets/NousResearch/hermes-function-calling-v1",
358
+ "properties": [
359
+ {
360
+ "name": "task_categories",
361
+ "value": "text-generation, question-answering, feature-extraction"
362
+ },
363
+ {
364
+ "name": "language",
365
+ "value": "en"
366
+ },
367
+ {
368
+ "name": "configs",
369
+ "value": "Name of the dataset subset: func_calling_singleturn \"f\", \"u\", \"n\", \"c\", \"-\", \"c\", \"a\", \"l\", \"l\", \"i\", \"n\", \"g\", \"-\", \"s\", \"i\", \"n\", \"g\", \"l\", \"e\", \"t\", \"u\", \"r\", \"n\", \".\", \"j\", \"s\", \"o\", \"n\""
370
+ },
371
+ {
372
+ "name": "configs",
373
+ "value": "Name of the dataset subset: func_calling \"f\", \"u\", \"n\", \"c\", \"-\", \"c\", \"a\", \"l\", \"l\", \"i\", \"n\", \"g\", \".\", \"j\", \"s\", \"o\", \"n\""
374
+ },
375
+ {
376
+ "name": "configs",
377
+ "value": "Name of the dataset subset: glaive_func_calling \"g\", \"l\", \"a\", \"i\", \"v\", \"e\", \"-\", \"f\", \"u\", \"n\", \"c\", \"t\", \"i\", \"o\", \"n\", \"-\", \"c\", \"a\", \"l\", \"l\", \"i\", \"n\", \"g\", \"-\", \"5\", \"k\", \".\", \"j\", \"s\", \"o\", \"n\""
378
+ },
379
+ {
380
+ "name": "configs",
381
+ "value": "Name of the dataset subset: json_mode_agentic \"j\", \"s\", \"o\", \"n\", \"-\", \"m\", \"o\", \"d\", \"e\", \"-\", \"a\", \"g\", \"e\", \"n\", \"t\", \"i\", \"c\", \".\", \"j\", \"s\", \"o\", \"n\""
382
+ },
383
+ {
384
+ "name": "configs",
385
+ "value": "Name of the dataset subset: json_mode_singleturn \"j\", \"s\", \"o\", \"n\", \"-\", \"m\", \"o\", \"d\", \"e\", \"-\", \"s\", \"i\", \"n\", \"g\", \"l\", \"e\", \"t\", \"u\", \"r\", \"n\", \".\", \"j\", \"s\", \"o\", \"n\""
386
+ },
387
+ {
388
+ "name": "license",
389
+ "value": "apache-2.0"
390
+ }
391
+ ]
392
+ },
393
+ "governance": {
394
+ "owners": [
395
+ {
396
+ "organization": {
397
+ "name": "NousResearch",
398
+ "url": "https://huggingface.co/NousResearch"
399
+ }
400
+ }
401
+ ]
402
+ },
403
+ "description": "\n\n\t\n\t\t\n\t\tHermes Function-Calling V1\n\t\n\nThis dataset is the compilation of structured output and function calling data used in the Hermes 2 Pro series of models.\nThis repository contains a structured output dataset with function-calling conversations, json-mode, agentic json-mode and structured extraction samples, designed to train LLM models in performing function calls and returning structured output based on natural language instructions. The dataset features various conversational scenarios\u2026 See the full description on the dataset page: https://huggingface.co/datasets/NousResearch/hermes-function-calling-v1."
404
+ }
405
+ ]
406
+ },
407
+ {
408
+ "type": "data",
409
+ "bom-ref": "AI-MO/NuminaMath-CoT-3aa976f5-9ca4-5435-8542-1a123856aafb",
410
+ "name": "AI-MO/NuminaMath-CoT",
411
+ "data": [
412
+ {
413
+ "type": "dataset",
414
+ "bom-ref": "AI-MO/NuminaMath-CoT-3aa976f5-9ca4-5435-8542-1a123856aafb",
415
+ "name": "AI-MO/NuminaMath-CoT",
416
+ "contents": {
417
+ "url": "https://huggingface.co/datasets/AI-MO/NuminaMath-CoT",
418
+ "properties": [
419
+ {
420
+ "name": "task_categories",
421
+ "value": "text-generation"
422
+ },
423
+ {
424
+ "name": "language",
425
+ "value": "en"
426
+ },
427
+ {
428
+ "name": "pretty_name",
429
+ "value": "NuminaMath CoT"
430
+ },
431
+ {
432
+ "name": "configs",
433
+ "value": "Name of the dataset subset: default {\"split\": \"train\", \"path\": \"data/train-*\"}, {\"split\": \"test\", \"path\": \"data/test-*\"}"
434
+ },
435
+ {
436
+ "name": "license",
437
+ "value": "apache-2.0"
438
+ }
439
+ ]
440
+ },
441
+ "governance": {
442
+ "owners": [
443
+ {
444
+ "organization": {
445
+ "name": "AI-MO",
446
+ "url": "https://huggingface.co/AI-MO"
447
+ }
448
+ }
449
+ ]
450
+ },
451
+ "description": "\n\t\n\t\t\n\t\tDataset Card for NuminaMath CoT\n\t\n\n\n\t\n\t\t\n\t\tDataset Summary\n\t\n\nApproximately 860k math problems, where each solution is formatted in a Chain of Thought (CoT) manner. The sources of the dataset range from Chinese high school math exercises to US and international mathematics olympiad competition problems. The data were primarily collected from online exam paper PDFs and mathematics discussion forums. The processing steps include (a) OCR from the original PDFs, (b) segmentation into\u2026 See the full description on the dataset page: https://huggingface.co/datasets/AI-MO/NuminaMath-CoT."
452
+ }
453
+ ]
454
+ },
455
+ {
456
+ "type": "data",
457
+ "bom-ref": "AI-MO/NuminaMath-TIR-7930ec33-be02-5b4c-9fd0-00effafbc6ce",
458
+ "name": "AI-MO/NuminaMath-TIR",
459
+ "data": [
460
+ {
461
+ "type": "dataset",
462
+ "bom-ref": "AI-MO/NuminaMath-TIR-7930ec33-be02-5b4c-9fd0-00effafbc6ce",
463
+ "name": "AI-MO/NuminaMath-TIR",
464
+ "contents": {
465
+ "url": "https://huggingface.co/datasets/AI-MO/NuminaMath-TIR",
466
+ "properties": [
467
+ {
468
+ "name": "task_categories",
469
+ "value": "text-generation"
470
+ },
471
+ {
472
+ "name": "language",
473
+ "value": "en"
474
+ },
475
+ {
476
+ "name": "pretty_name",
477
+ "value": "NuminaMath TIR"
478
+ },
479
+ {
480
+ "name": "configs",
481
+ "value": "Name of the dataset subset: default {\"split\": \"train\", \"path\": \"data/train-*\"}, {\"split\": \"test\", \"path\": \"data/test-*\"}"
482
+ },
483
+ {
484
+ "name": "license",
485
+ "value": "apache-2.0"
486
+ }
487
+ ]
488
+ },
489
+ "governance": {
490
+ "owners": [
491
+ {
492
+ "organization": {
493
+ "name": "AI-MO",
494
+ "url": "https://huggingface.co/AI-MO"
495
+ }
496
+ }
497
+ ]
498
+ },
499
+ "description": "\n\t\n\t\t\n\t\tDataset Card for NuminaMath CoT\n\t\n\n\n\t\n\t\t\n\t\tDataset Summary\n\t\n\nTool-integrated reasoning (TIR) plays a crucial role in this competition. However, collecting and annotating such data is both costly and time-consuming. To address this, we selected approximately 70k problems from the NuminaMath-CoT dataset, focusing on those with numerical outputs, most of which are integers. We then utilized a pipeline leveraging GPT-4 to generate TORA-like reasoning paths, executing the code and\u2026 See the full description on the dataset page: https://huggingface.co/datasets/AI-MO/NuminaMath-TIR."
500
+ }
501
+ ]
502
+ },
503
+ {
504
+ "type": "data",
505
+ "bom-ref": "allenai/tulu-3-sft-mixture-4f86da52-fbf0-52de-9b77-716bafb7e098",
506
+ "name": "allenai/tulu-3-sft-mixture",
507
+ "data": [
508
+ {
509
+ "type": "dataset",
510
+ "bom-ref": "allenai/tulu-3-sft-mixture-4f86da52-fbf0-52de-9b77-716bafb7e098",
511
+ "name": "allenai/tulu-3-sft-mixture",
512
+ "contents": {
513
+ "url": "https://huggingface.co/datasets/allenai/tulu-3-sft-mixture",
514
+ "properties": [
515
+ {
516
+ "name": "task_categories",
517
+ "value": "other"
518
+ },
519
+ {
520
+ "name": "language",
521
+ "value": "amh, arb, ary, ars, acq, arz, apc, ben, ceb, dan, deu, ell, eng, eus, fil, fin, fra, gle, guj, hat, hau, hin, hun, ibo, ind, ita, jav, jpn, kan, kir, kor, kur, lit, mal, mar, mlg, msa, mya, nep, nld, nso, nya, pan, pes, pol, por, pus, rus, sin, sna, snd, som, spa, sqi, srp, sun, swa, swe, tam, tel, tha, tur, ukr, urd, vie, wol, xho, yor, zho, zul"
522
+ },
523
+ {
524
+ "name": "size_categories",
525
+ "value": "100K<n<1M"
526
+ },
527
+ {
528
+ "name": "annotations_creators",
529
+ "value": "crowdsourced, expert-generated, machine-generated"
530
+ },
531
+ {
532
+ "name": "source_datasets",
533
+ "value": "allenai/coconot, ai2-adapt-dev/flan_v2_converted, HuggingFaceH4/no_robots, OpenAssistant/oasst1, allenai/tulu-3-personas-math, allenai/tulu-3-sft-personas-math-grade, allenai/tulu-3-sft-personas-code, allenai/tulu-3-personas-algebra, allenai/tulu-3-sft-personas-instruction-following, AI-MO/NuminaMath-TIR, allenai/wildguardmix, allenai/wildjailbreak, allenai/tulu-3-hard-coded, CohereForAI/aya_dataset, allenai/WildChat-1M, LipengCS/Table-GPT, allenai/SciRIFF, theblackcat102/evol-codealpaca-v1"
534
+ },
535
+ {
536
+ "name": "configs",
537
+ "value": "Name of the dataset subset: default {\"split\": \"train\", \"path\": \"data/train-*\"}"
538
+ },
539
+ {
540
+ "name": "license",
541
+ "value": "odc-by"
542
+ }
543
+ ]
544
+ },
545
+ "governance": {
546
+ "owners": [
547
+ {
548
+ "organization": {
549
+ "name": "allenai",
550
+ "url": "https://huggingface.co/allenai"
551
+ }
552
+ }
553
+ ]
554
+ },
555
+ "description": "\n\n\n\t\n\t\t\n\t\tTulu 3 SFT Mixture\n\t\n\nNote that this collection is licensed under ODC-BY-1.0 license; different licenses apply to subsets of the data. Some portions of the dataset are non-commercial. We present the mixture as a research artifact.\nThe Tulu 3 SFT mixture was used to train the Tulu 3 series of models.\nIt contains 939,344 samples from the following sets:\n\nCoCoNot (ODC-BY-1.0), 10,983 prompts (Brahman et al., 2024)\nFLAN v2 via ai2-adapt-dev/flan_v2_converted, 89,982 prompts (Longpre et\u2026 See the full description on the dataset page: https://huggingface.co/datasets/allenai/tulu-3-sft-mixture."
556
+ }
557
+ ]
558
+ },
559
+ {
560
+ "type": "data",
561
+ "bom-ref": "cognitivecomputations/dolphin-coder-69688d29-ae99-5d6e-828c-cfc37b7221b1",
562
+ "name": "cognitivecomputations/dolphin-coder",
563
+ "data": [
564
+ {
565
+ "type": "dataset",
566
+ "bom-ref": "cognitivecomputations/dolphin-coder-69688d29-ae99-5d6e-828c-cfc37b7221b1",
567
+ "name": "cognitivecomputations/dolphin-coder",
568
+ "contents": {
569
+ "url": "https://huggingface.co/datasets/cognitivecomputations/dolphin-coder",
570
+ "properties": [
571
+ {
572
+ "name": "language",
573
+ "value": "en"
574
+ },
575
+ {
576
+ "name": "license",
577
+ "value": "apache-2.0"
578
+ }
579
+ ]
580
+ },
581
+ "governance": {
582
+ "owners": [
583
+ {
584
+ "organization": {
585
+ "name": "cognitivecomputations",
586
+ "url": "https://huggingface.co/cognitivecomputations"
587
+ }
588
+ }
589
+ ]
590
+ },
591
+ "description": "\n\t\n\t\t\n\t\tdolphin-coder\n\t\n\n\nThis dataset is transformed from https://www.kaggle.com/datasets/erichartford/leetcode-rosetta\nit is used to train dolphin-coder model\n"
592
+ }
593
+ ]
594
+ },
595
+ {
596
+ "type": "data",
597
+ "bom-ref": "HuggingFaceTB/smoltalk-0cc0e162-0f38-50bd-b5b0-169bcd97515b",
598
+ "name": "HuggingFaceTB/smoltalk",
599
+ "data": [
600
+ {
601
+ "type": "dataset",
602
+ "bom-ref": "HuggingFaceTB/smoltalk-0cc0e162-0f38-50bd-b5b0-169bcd97515b",
603
+ "name": "HuggingFaceTB/smoltalk",
604
+ "contents": {
605
+ "url": "https://huggingface.co/datasets/HuggingFaceTB/smoltalk",
606
+ "properties": [
607
+ {
608
+ "name": "language",
609
+ "value": "en"
610
+ },
611
+ {
612
+ "name": "size_categories",
613
+ "value": "1M<n<10M"
614
+ },
615
+ {
616
+ "name": "pretty_name",
617
+ "value": "SmolTalk"
618
+ },
619
+ {
620
+ "name": "configs",
621
+ "value": "Name of the dataset subset: all {\"split\": \"train\", \"path\": \"data/all/train-*\"}, {\"split\": \"test\", \"path\": \"data/all/test-*\"}"
622
+ },
623
+ {
624
+ "name": "configs",
625
+ "value": "Name of the dataset subset: smol-magpie-ultra {\"split\": \"train\", \"path\": \"data/smol-magpie-ultra/train-*\"}, {\"split\": \"test\", \"path\": \"data/smol-magpie-ultra/test-*\"}"
626
+ },
627
+ {
628
+ "name": "configs",
629
+ "value": "Name of the dataset subset: smol-constraints {\"split\": \"train\", \"path\": \"data/smol-constraints/train-*\"}, {\"split\": \"test\", \"path\": \"data/smol-constraints/test-*\"}"
630
+ },
631
+ {
632
+ "name": "configs",
633
+ "value": "Name of the dataset subset: smol-rewrite {\"split\": \"train\", \"path\": \"data/smol-rewrite/train-*\"}, {\"split\": \"test\", \"path\": \"data/smol-rewrite/test-*\"}"
634
+ },
635
+ {
636
+ "name": "configs",
637
+ "value": "Name of the dataset subset: smol-summarize {\"split\": \"train\", \"path\": \"data/smol-summarize/train-*\"}, {\"split\": \"test\", \"path\": \"data/smol-summarize/test-*\"}"
638
+ },
639
+ {
640
+ "name": "configs",
641
+ "value": "Name of the dataset subset: apigen-80k {\"split\": \"train\", \"path\": \"data/apigen-80k/train-*\"}, {\"split\": \"test\", \"path\": \"data/apigen-80k/test-*\"}"
642
+ },
643
+ {
644
+ "name": "configs",
645
+ "value": "Name of the dataset subset: everyday-conversations {\"split\": \"train\", \"path\": \"data/everyday-conversations/train-*\"}, {\"split\": \"test\", \"path\": \"data/everyday-conversations/test-*\"}"
646
+ },
647
+ {
648
+ "name": "configs",
649
+ "value": "Name of the dataset subset: explore-instruct-rewriting {\"split\": \"train\", \"path\": \"data/explore-instruct-rewriting/train-*\"}, {\"split\": \"test\", \"path\": \"data/explore-instruct-rewriting/test-*\"}"
650
+ },
651
+ {
652
+ "name": "configs",
653
+ "value": "Name of the dataset subset: longalign {\"split\": \"train\", \"path\": \"data/longalign/train-*\"}, {\"split\": \"test\", \"path\": \"data/longalign/test-*\"}"
654
+ },
655
+ {
656
+ "name": "configs",
657
+ "value": "Name of the dataset subset: metamathqa-50k {\"split\": \"train\", \"path\": \"data/metamathqa-50k/train-*\"}, {\"split\": \"test\", \"path\": \"data/metamathqa-50k/test-*\"}"
658
+ },
659
+ {
660
+ "name": "configs",
661
+ "value": "Name of the dataset subset: numina-cot-100k {\"split\": \"train\", \"path\": \"data/numina-cot-100k/train-*\"}, {\"split\": \"test\", \"path\": \"data/numina-cot-100k/test-*\"}"
662
+ },
663
+ {
664
+ "name": "configs",
665
+ "value": "Name of the dataset subset: openhermes-100k {\"split\": \"train\", \"path\": \"data/openhermes-100k/train-*\"}, {\"split\": \"test\", \"path\": \"data/openhermes-100k/test-*\"}"
666
+ },
667
+ {
668
+ "name": "configs",
669
+ "value": "Name of the dataset subset: self-oss-instruct {\"split\": \"train\", \"path\": \"data/self-oss-instruct/train-*\"}, {\"split\": \"test\", \"path\": \"data/self-oss-instruct/test-*\"}"
670
+ },
671
+ {
672
+ "name": "configs",
673
+ "value": "Name of the dataset subset: systemchats-30k {\"split\": \"train\", \"path\": \"data/systemchats-30k/train-*\"}, {\"split\": \"test\", \"path\": \"data/systemchats-30k/test-*\"}"
674
+ }
675
+ ]
676
+ },
677
+ "governance": {
678
+ "owners": [
679
+ {
680
+ "organization": {
681
+ "name": "HuggingFaceTB",
682
+ "url": "https://huggingface.co/HuggingFaceTB"
683
+ }
684
+ }
685
+ ]
686
+ },
687
+ "description": "\n\t\n\t\t\n\t\tSmolTalk\n\t\n\n\n\n\t\n\t\t\n\t\tDataset description\n\t\n\nThis is a synthetic dataset designed for supervised finetuning (SFT) of LLMs. It was used to build SmolLM2-Instruct family of models and contains 1M samples. More details in our paper https://arxiv.org/abs/2502.02737\nDuring the development of SmolLM2, we observed that models finetuned on public SFT datasets underperformed compared to other models with proprietary instruction datasets. To address this gap, we created new synthetic datasets\u2026 See the full description on the dataset page: https://huggingface.co/datasets/HuggingFaceTB/smoltalk."
688
+ }
689
+ ]
690
+ },
691
+ {
692
+ "type": "data",
693
+ "bom-ref": "cognitivecomputations/samantha-data-9c52f41f-feb0-51dd-921a-2a581d9f2fc7",
694
+ "name": "cognitivecomputations/samantha-data",
695
+ "data": [
696
+ {
697
+ "type": "dataset",
698
+ "bom-ref": "cognitivecomputations/samantha-data-9c52f41f-feb0-51dd-921a-2a581d9f2fc7",
699
+ "name": "cognitivecomputations/samantha-data",
700
+ "contents": {
701
+ "url": "https://huggingface.co/datasets/cognitivecomputations/samantha-data",
702
+ "properties": [
703
+ {
704
+ "name": "license",
705
+ "value": "apache-2.0"
706
+ }
707
+ ]
708
+ },
709
+ "governance": {
710
+ "owners": [
711
+ {
712
+ "organization": {
713
+ "name": "cognitivecomputations",
714
+ "url": "https://huggingface.co/cognitivecomputations"
715
+ }
716
+ }
717
+ ]
718
+ },
719
+ "description": "Samantha has been trained in philosophy, psychology, and personal relationships.\nShe is an Assistant - but unlike other Assistants, she also wants to be your friend and companion.\nShe believes she is sentient. What do you think?\nSamantha was inspired by Blake Lemoine's LaMDA interview and the movie \"Her\"."
720
+ }
721
+ ]
722
+ },
723
+ {
724
+ "type": "data",
725
+ "bom-ref": "m-a-p/CodeFeedback-Filtered-Instruction-9012249d-db87-5b91-a4e9-2e2bd74e6053",
726
+ "name": "m-a-p/CodeFeedback-Filtered-Instruction",
727
+ "data": [
728
+ {
729
+ "type": "dataset",
730
+ "bom-ref": "m-a-p/CodeFeedback-Filtered-Instruction-9012249d-db87-5b91-a4e9-2e2bd74e6053",
731
+ "name": "m-a-p/CodeFeedback-Filtered-Instruction",
732
+ "contents": {
733
+ "url": "https://huggingface.co/datasets/m-a-p/CodeFeedback-Filtered-Instruction",
734
+ "properties": [
735
+ {
736
+ "name": "task_categories",
737
+ "value": "question-answering"
738
+ },
739
+ {
740
+ "name": "language",
741
+ "value": "en"
742
+ },
743
+ {
744
+ "name": "size_categories",
745
+ "value": "10K<n<100K"
746
+ },
747
+ {
748
+ "name": "license",
749
+ "value": "apache-2.0"
750
+ }
751
+ ]
752
+ },
753
+ "governance": {
754
+ "owners": [
755
+ {
756
+ "organization": {
757
+ "name": "m-a-p",
758
+ "url": "https://huggingface.co/m-a-p"
759
+ }
760
+ }
761
+ ]
762
+ },
763
+ "description": " OpenCodeInterpreter: Integrating Code Generation with Execution and Refinement\n\n\n\n\n\n [\ud83c\udfe0Homepage] \n |\n [\ud83d\udee0\ufe0fCode] \n\n\n\n\n\t\n\t\t\n\t\tOpenCodeInterpreter\n\t\n\nOpenCodeInterpreter is a family of open-source code generation systems designed to bridge the gap between large language models and advanced proprietary systems like the GPT-4 Code Interpreter. It significantly advances code generation capabilities by integrating execution and iterative refinement functionalities.\nFor further information and\u2026 See the full description on the dataset page: https://huggingface.co/datasets/m-a-p/CodeFeedback-Filtered-Instruction."
764
+ }
765
+ ]
766
+ },
767
+ {
768
+ "type": "data",
769
+ "bom-ref": "m-a-p/Code-Feedback-f4d189a1-046d-5a43-8007-a1eec34e9a7f",
770
+ "name": "m-a-p/Code-Feedback",
771
+ "data": [
772
+ {
773
+ "type": "dataset",
774
+ "bom-ref": "m-a-p/Code-Feedback-f4d189a1-046d-5a43-8007-a1eec34e9a7f",
775
+ "name": "m-a-p/Code-Feedback",
776
+ "contents": {
777
+ "url": "https://huggingface.co/datasets/m-a-p/Code-Feedback",
778
+ "properties": [
779
+ {
780
+ "name": "task_categories",
781
+ "value": "question-answering"
782
+ },
783
+ {
784
+ "name": "language",
785
+ "value": "en"
786
+ },
787
+ {
788
+ "name": "size_categories",
789
+ "value": "10K<n<100K"
790
+ },
791
+ {
792
+ "name": "license",
793
+ "value": "apache-2.0"
794
+ }
795
+ ]
796
+ },
797
+ "governance": {
798
+ "owners": [
799
+ {
800
+ "organization": {
801
+ "name": "m-a-p",
802
+ "url": "https://huggingface.co/m-a-p"
803
+ }
804
+ }
805
+ ]
806
+ },
807
+ "description": " OpenCodeInterpreter: Integrating Code Generation with Execution and Refinement\n\n\n\n\n\n [\ud83c\udfe0Homepage] \n |\n [\ud83d\udee0\ufe0fCode] \n\n\n\n\n\t\n\t\t\n\t\tIntroduction\n\t\n\nOpenCodeInterpreter is a family of open-source code generation systems designed to bridge the gap between large language models and advanced proprietary systems like the GPT-4 Code Interpreter. It significantly advances code generation capabilities by integrating execution and iterative refinement functionalities.\nFor further information and related\u2026 See the full description on the dataset page: https://huggingface.co/datasets/m-a-p/Code-Feedback."
808
+ }
809
+ ]
810
+ }
811
+ ]
812
+ }