|
|
--- |
|
|
dataset_info: |
|
|
features: |
|
|
- name: instruction |
|
|
dtype: string |
|
|
- name: context |
|
|
dtype: string |
|
|
- name: response |
|
|
dtype: string |
|
|
- name: category |
|
|
dtype: string |
|
|
splits: |
|
|
- name: train |
|
|
num_bytes: 3243541 |
|
|
num_examples: 4000 |
|
|
download_size: 2050955 |
|
|
dataset_size: 3243541 |
|
|
configs: |
|
|
- config_name: default |
|
|
data_files: |
|
|
- split: train |
|
|
path: data/train-* |
|
|
license: apache-2.0 |
|
|
task_categories: |
|
|
- table-question-answering |
|
|
- question-answering |
|
|
- text-generation |
|
|
language: |
|
|
- en |
|
|
size_categories: |
|
|
- 1K<n<10K |
|
|
--- |
|
|
|
|
|
|
|
|
# Databricks-Dolly-8k |
|
|
|
|
|
The resulting dataset contains **8000 samples** of the [databricks/databricks-dolly-15k](https://huggingface.co/datasets/databricks/databricks-dolly-15k) dataset. |
|
|
|
|
|
This split of an even smaller subset is provided for very fast experimentation and evaluation of models when computational resources are highly limited or for quick prototyping. |
|
|
|
|
|
## Dataset Structure |
|
|
|
|
|
The dataset is provided as a `DatasetDict` with the following splits: |
|
|
|
|
|
* **`train`**: Contains 8000 samples. |
|
|
|
|
|
Each split contains the following features, identical to the original dataset: |
|
|
|
|
|
* `id`: The unique identifier for each sample. |
|
|
* `instruction`: The instruction or prompt for the task. |
|
|
* `response`: The response to the given instruction. |
|
|
* `context`: Additional context or information related to the instruction. |
|
|
* `source`: The source of the sample. |
|
|
|
|
|
## Usage |
|
|
|
|
|
You can easily load this split dataset using the `datasets` library: |
|
|
|
|
|
```python |
|
|
from datasets import load_dataset |
|
|
|
|
|
databricks_dolly_8k = load_dataset("Vishva007/Databricks-Dolly-8k") |
|
|
|
|
|
print(databricks_dolly_8k) |
|
|
print(databricks_dolly_8k["train"][0]) |
|
|
``` |
|
|
|
|
|
## Example Usage |
|
|
|
|
|
Here’s an example of how you might use this dataset in a Python script: |
|
|
|
|
|
```python |
|
|
from datasets import load_dataset |
|
|
|
|
|
# Load the dataset |
|
|
databricks_dolly_8k = load_dataset("Vishva007/Databricks-Dolly-8k") |
|
|
|
|
|
# Print the first sample in the training set |
|
|
print(databricks_dolly_8k["train"][0]) |
|
|
|
|
|
# Access specific fields from the first sample |
|
|
sample = databricks_dolly_8k["train"][0] |
|
|
print(f"ID: {sample['id']}") |
|
|
print(f"Instruction: {sample['instruction']}") |
|
|
print(f"Response: {sample['response']}") |
|
|
print(f"Context: {sample['context']}") |
|
|
print(f"Source: {sample['source']}") |
|
|
``` |
|
|
|
|
|
## Dataset Info |
|
|
|
|
|
### Features |
|
|
|
|
|
- `id`: The unique identifier for each sample. |
|
|
- `instruction`: The instruction or prompt for the task. |
|
|
- `response`: The response to the given instruction. |
|
|
- `context`: Additional context or information related to the instruction. |
|
|
- `source`: The source of the sample. |
|
|
|
|
|
### Splits |
|
|
|
|
|
- **`train`**: Contains 8000 samples. |
|
|
|
|
|
## License |
|
|
|
|
|
This dataset is derived from the [databricks/databricks-dolly-15k](https://huggingface.co/datasets/databricks/databricks-dolly-15k) dataset, which is licensed under the [Apache License 2.0](https://www.apache.org/licenses/LICENSE-2.0). |
|
|
|
|
|
For more details about the original dataset, please refer to the [official documentation](https://huggingface.co/datasets/databricks/databricks-dolly-15k). |
|
|
|
|
|
--- |