File size: 1,870 Bytes
b380e9a f2876f5 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 |
---
dataset_info:
features:
- name: document
dtype: string
- name: summary
dtype: string
splits:
- name: train
num_bytes: 23380
num_examples: 100
- name: validation
num_bytes: 23634
num_examples: 100
- name: test
num_bytes: 24038
num_examples: 100
download_size: 55197
dataset_size: 71052
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
- split: test
path: data/test-*
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
This is a tiny version of [https://huggingface.co/datasets/Harvard/gigaword](Harvard/gigaword), used for testing purposes.
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
This is a tiny version of [https://huggingface.co/datasets/Harvard/gigaword](Harvard/gigaword).
It was created by selecting only first 100 samples from each split.
- **Language(s) (NLP):** English
### Dataset Sources [optional]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
It is supposed be used only for testing purposes.
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
It is supposed be used only when you want to test that your code works.
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
Dataset contains two string columns: `document` and `summary`. `document` shows source document, and `summary` is summarization of document.
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
To test code without waiting a long time to download gigawords. |