ianomunga commited on
Commit
b3e7300
·
verified ·
1 Parent(s): e65f181

AfriCompute aggregated dataset: dolly

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -5
  2. README.md +116 -0
  3. databricks-dolly-15k.jsonl +3 -0
.gitattributes CHANGED
@@ -9,7 +9,6 @@
9
  *.joblib filter=lfs diff=lfs merge=lfs -text
10
  *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
  *.lz4 filter=lfs diff=lfs merge=lfs -text
12
- *.mds filter=lfs diff=lfs merge=lfs -text
13
  *.mlmodel filter=lfs diff=lfs merge=lfs -text
14
  *.model filter=lfs diff=lfs merge=lfs -text
15
  *.msgpack filter=lfs diff=lfs merge=lfs -text
@@ -27,7 +26,6 @@
27
  *.safetensors filter=lfs diff=lfs merge=lfs -text
28
  saved_model/**/* filter=lfs diff=lfs merge=lfs -text
29
  *.tar.* filter=lfs diff=lfs merge=lfs -text
30
- *.tar filter=lfs diff=lfs merge=lfs -text
31
  *.tflite filter=lfs diff=lfs merge=lfs -text
32
  *.tgz filter=lfs diff=lfs merge=lfs -text
33
  *.wasm filter=lfs diff=lfs merge=lfs -text
@@ -54,6 +52,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
54
  *.jpg filter=lfs diff=lfs merge=lfs -text
55
  *.jpeg filter=lfs diff=lfs merge=lfs -text
56
  *.webp filter=lfs diff=lfs merge=lfs -text
57
- # Video files - compressed
58
- *.mp4 filter=lfs diff=lfs merge=lfs -text
59
- *.webm filter=lfs diff=lfs merge=lfs -text
 
9
  *.joblib filter=lfs diff=lfs merge=lfs -text
10
  *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
  *.lz4 filter=lfs diff=lfs merge=lfs -text
 
12
  *.mlmodel filter=lfs diff=lfs merge=lfs -text
13
  *.model filter=lfs diff=lfs merge=lfs -text
14
  *.msgpack filter=lfs diff=lfs merge=lfs -text
 
26
  *.safetensors filter=lfs diff=lfs merge=lfs -text
27
  saved_model/**/* filter=lfs diff=lfs merge=lfs -text
28
  *.tar.* filter=lfs diff=lfs merge=lfs -text
 
29
  *.tflite filter=lfs diff=lfs merge=lfs -text
30
  *.tgz filter=lfs diff=lfs merge=lfs -text
31
  *.wasm filter=lfs diff=lfs merge=lfs -text
 
52
  *.jpg filter=lfs diff=lfs merge=lfs -text
53
  *.jpeg filter=lfs diff=lfs merge=lfs -text
54
  *.webp filter=lfs diff=lfs merge=lfs -text
55
+ databricks-dolly-15k.jsonl filter=lfs diff=lfs merge=lfs -text
 
 
README.md ADDED
@@ -0,0 +1,116 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-sa-3.0
3
+ task_categories:
4
+ - question-answering
5
+ - summarization
6
+ language:
7
+ - en
8
+ size_categories:
9
+ - 10K<n<100K
10
+ ---
11
+ # Summary
12
+ `databricks-dolly-15k` is an open source dataset of instruction-following records generated by thousands of Databricks employees in several
13
+ of the behavioral categories outlined in the [InstructGPT](https://arxiv.org/abs/2203.02155) paper, including brainstorming, classification,
14
+ closed QA, generation, information extraction, open QA, and summarization.
15
+
16
+ This dataset can be used for any purpose, whether academic or commercial, under the terms of the
17
+ [Creative Commons Attribution-ShareAlike 3.0 Unported License](https://creativecommons.org/licenses/by-sa/3.0/legalcode).
18
+
19
+ Supported Tasks:
20
+ - Training LLMs
21
+ - Synthetic Data Generation
22
+ - Data Augmentation
23
+
24
+ Languages: English
25
+ Version: 1.0
26
+
27
+ **Owner: Databricks, Inc.**
28
+
29
+
30
+ # Dataset Overview
31
+ `databricks-dolly-15k` is a corpus of more than 15,000 records generated by thousands of Databricks employees to enable large language
32
+ models to exhibit the magical interactivity of ChatGPT.
33
+ Databricks employees were invited to create prompt / response pairs in each of eight different instruction categories, including
34
+ the seven outlined in the InstructGPT paper, as well as an open-ended free-form category. The contributors were instructed to avoid using
35
+ information from any source on the web with the exception of Wikipedia (for particular subsets of instruction categories), and explicitly
36
+ instructed to avoid using generative AI in formulating instructions or responses. Examples of each behavior were provided to motivate the
37
+ types of questions and instructions appropriate to each category.
38
+
39
+ Halfway through the data generation process, contributors were given the option of answering questions posed by other contributors.
40
+ They were asked to rephrase the original question and only select questions they could be reasonably expected to answer correctly.
41
+
42
+ For certain categories contributors were asked to provide reference texts copied from Wikipedia. Reference text (indicated by the `context`
43
+ field in the actual dataset) may contain bracketed Wikipedia citation numbers (e.g. `[42]`) which we recommend users remove for downstream applications.
44
+
45
+
46
+ # Intended Uses
47
+ While immediately valuable for instruction fine tuning large language models, as a corpus of human-generated instruction prompts,
48
+ this dataset also presents a valuable opportunity for synthetic data generation in the methods outlined in the Self-Instruct paper.
49
+ For example, contributor--generated prompts could be submitted as few-shot examples to a large open language model to generate a
50
+ corpus of millions of examples of instructions in each of the respective InstructGPT categories.
51
+
52
+ Likewise, both the instructions and responses present fertile ground for data augmentation. A paraphrasing model might be used to
53
+ restate each prompt or short responses, with the resulting text associated to the respective ground-truth sample. Such an approach might
54
+ provide a form of regularization on the dataset that could allow for more robust instruction-following behavior in models derived from
55
+ these synthetic datasets.
56
+
57
+
58
+ # Dataset
59
+ ## Purpose of Collection
60
+ As part of our continuing commitment to open source, Databricks developed what is, to the best of our knowledge, the first open source,
61
+ human-generated instruction corpus specifically designed to enable large language models to exhibit the magical interactivity of ChatGPT.
62
+ Unlike other datasets that are limited to non-commercial use, this dataset can be used, modified, and extended for any purpose, including
63
+ academic or commercial applications.
64
+
65
+ ## Sources
66
+ - **Human-generated data**: Databricks employees were invited to create prompt / response pairs in each of eight different instruction categories.
67
+ - **Wikipedia**: For instruction categories that require an annotator to consult a reference text (information extraction, closed QA, summarization)
68
+ contributors selected passages from Wikipedia for particular subsets of instruction categories. No guidance was given to annotators as to how to select the
69
+ target passages.
70
+
71
+ ## Annotator Guidelines
72
+ To create a record, employees were given a brief description of the annotation task as well as examples of the types of prompts typical
73
+ of each annotation task. Guidelines were succinct by design so as to encourage a high task completion rate, possibly at the cost of
74
+ rigorous compliance to an annotation rubric that concretely and reliably operationalizes the specific task. Caveat emptor.
75
+
76
+ The annotation guidelines for each of the categories are as follows:
77
+
78
+ - **Creative Writing**: Write a question or instruction that requires a creative, open-ended written response. The instruction should be reasonable to ask of a person with general world knowledge and should not require searching. In this task, your prompt should give very specific instructions to follow. Constraints, instructions, guidelines, or requirements all work, and the more of them the better.
79
+ - **Closed QA**: Write a question or instruction that requires factually correct response based on a passage of text from Wikipedia. The question can be complex and can involve human-level reasoning capabilities, but should not require special knowledge. To create a question for this task include both the text of the question as well as the reference text in the form.
80
+ - **Open QA**: Write a question that can be answered using general world knowledge or at most a single search. This task asks for opinions and facts about the world at large and does not provide any reference text for consultation.
81
+ - **Summarization**: Give a summary of a paragraph from Wikipedia. Please don't ask questions that will require more than 3-5 minutes to answer. To create a question for this task include both the text of the question as well as the reference text in the form.
82
+ - **Information Extraction**: These questions involve reading a paragraph from Wikipedia and extracting information from the passage. Everything required to produce an answer (e.g. a list, keywords etc) should be included in the passages. To create a question for this task include both the text of the question as well as the reference text in the form.
83
+ - **Classification**: These prompts contain lists or examples of entities to be classified, e.g. movie reviews, products, etc. In this task the text or list of entities under consideration is contained in the prompt (e.g. there is no reference text.). You can choose any categories for classification you like, the more diverse the better.
84
+ - **Brainstorming**: Think up lots of examples in response to a question asking to brainstorm ideas.
85
+
86
+ ## Personal or Sensitive Data
87
+ This dataset contains public information (e.g., some information from Wikipedia). To our knowledge, there are no private person’s personal identifiers or sensitive information.
88
+
89
+ ## Language
90
+ American English
91
+
92
+ # Known Limitations
93
+ - Wikipedia is a crowdsourced corpus and the contents of this dataset may reflect the bias, factual errors and topical focus found in Wikipedia
94
+ - Some annotators may not be native English speakers
95
+ - Annotator demographics and subject matter may reflect the makeup of Databricks employees
96
+
97
+ # Citation
98
+
99
+ ```
100
+ @online{DatabricksBlog2023DollyV2,
101
+ author = {Mike Conover and Matt Hayes and Ankit Mathur and Jianwei Xie and Jun Wan and Sam Shah and Ali Ghodsi and Patrick Wendell and Matei Zaharia and Reynold Xin},
102
+ title = {Free Dolly: Introducing the World's First Truly Open Instruction-Tuned LLM},
103
+ year = {2023},
104
+ url = {https://www.databricks.com/blog/2023/04/12/dolly-first-open-commercially-viable-instruction-tuned-llm},
105
+ urldate = {2023-06-30}
106
+ }
107
+ ```
108
+
109
+ # License/Attribution
110
+ **Copyright (2023) Databricks, Inc.**
111
+ This dataset was developed at Databricks (https://www.databricks.com) and its use is subject to the CC BY-SA 3.0 license.
112
+
113
+ Certain categories of material in the dataset include materials from the following sources, licensed under the CC BY-SA 3.0 license:
114
+
115
+ Wikipedia (various pages) - https://www.wikipedia.org/
116
+ Copyright © Wikipedia editors and contributors.
databricks-dolly-15k.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2df9083338b4abd6bceb5635764dab5d833b393b55759dffb0959b6fcbf794ec
3
+ size 13085339