File size: 5,321 Bytes
5073e57 a5d60a7 5073e57 86d9f46 d1a1d66 e9a5685 a5d60a7 0c07c5f a5d60a7 0c07c5f a5d60a7 df83866 0c07c5f df83866 f708f8a df83866 ddb0a32 5073e57 2141783 a5d60a7 df83866 a5d60a7 ddb0a32 5073e57 fc73f1d 254cf03 fc73f1d 95bb4b4 fc73f1d 3175e80 fc73f1d 3175e80 fc73f1d 3175e80 fc73f1d 3175e80 fc73f1d 3175e80 fc73f1d 95bb4b4 fc73f1d 3175e80 fc73f1d 95bb4b4 fc73f1d 3175e80 95bb4b4 3175e80 95bb4b4 fc73f1d 3175e80 fc73f1d 3175e80 fc73f1d 3175e80 fc73f1d 95bb4b4 fc73f1d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 |
---
dataset_info:
- config_name: default
features:
- name: utterance
dtype: string
- name: label
sequence: int64
splits:
- name: train
num_bytes: 8999208
num_examples: 2742
- name: test
num_bytes: 1255307
num_examples: 378
download_size: 22576550
dataset_size: 10254515
- config_name: intents
features:
- name: id
dtype: int64
- name: name
dtype: string
- name: tags
sequence: 'null'
- name: regex_full_match
sequence: 'null'
- name: regex_partial_match
sequence: 'null'
- name: description
dtype: 'null'
splits:
- name: full_intents
num_bytes: 1240
num_examples: 29
- name: intents
num_bytes: 907
num_examples: 21
download_size: 8042
dataset_size: 2147
- config_name: intentsqwen3-32b
features:
- name: id
dtype: int64
- name: name
dtype: string
- name: tags
sequence: 'null'
- name: regex_full_match
sequence: 'null'
- name: regex_partial_match
sequence: 'null'
- name: description
dtype: string
splits:
- name: intents
num_bytes: 2497
num_examples: 21
download_size: 5062
dataset_size: 2497
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
- config_name: intents
data_files:
- split: full_intents
path: intents/full_intents-*
- split: intents
path: intents/intents-*
- config_name: intentsqwen3-32b
data_files:
- split: intents
path: intentsqwen3-32b/intents-*
---
# events
This is a text classification dataset. It is intended for machine learning research and experimentation.
This dataset is obtained via formatting another publicly available data to be compatible with our [AutoIntent Library](https://deeppavlov.github.io/AutoIntent/index.html).
## Usage
It is intended to be used with our [AutoIntent Library](https://deeppavlov.github.io/AutoIntent/index.html):
```python
from autointent import Dataset
banking77 = Dataset.from_hub("AutoIntent/events")
```
## Source
This dataset is taken from `knowledgator/events_classification_biotech` and formatted with our [AutoIntent Library](https://deeppavlov.github.io/AutoIntent/index.html):
```python
"""Convert events dataset to autointent internal format and scheme."""
from datasets import Dataset as HFDataset
from datasets import load_dataset
from autointent import Dataset
from autointent.schemas import Intent
def extract_intents_data(events_dataset: HFDataset) -> list[Intent]:
"""Extract intent names and assign ids to them."""
intent_names = sorted({name for intents in events_dataset["train"]["all_labels"] for name in intents})
return [Intent(id=i,name=name) for i, name in enumerate(intent_names)]
def converting_mapping(example: dict, intents_data: list[Intent]) -> dict[str, str | list[int] | None]:
"""Extract utterance and OHE label and drop the rest."""
res = {
"utterance": example["content"],
"label": [
int(intent.name in example["all_labels"]) for intent in intents_data
]
}
if sum(res["label"]) == 0:
res["label"] = None
return res
def convert_events(events_split: HFDataset, intents_data: dict[str, int]) -> list[dict]:
"""Convert one split into desired format."""
events_split = events_split.map(
converting_mapping, remove_columns=events_split.features.keys(),
fn_kwargs={"intents_data": intents_data}
)
return [sample for sample in events_split if sample["utterance"] is not None]
def get_low_resource_classes_mask(ds: list[dict], intent_names: list[str], fraction_thresh: float = 0.01) -> list[bool]:
res = [0] * len(intent_names)
for sample in ds:
for i, indicator in enumerate(sample["label"]):
res[i] += indicator
for i in range(len(intent_names)):
res[i] /= len(ds)
return [(frac < fraction_thresh) for frac in res]
def remove_low_resource_classes(ds: list[dict], mask: list[bool]) -> list[dict]:
res = []
for sample in ds:
if sum(sample["label"]) == 1 and mask[sample["label"].index(1)]:
continue
sample["label"] = [
indicator for indicator, low_resource in
zip(sample["label"], mask, strict=True) if not low_resource
]
res.append(sample)
return res
def remove_oos(ds: list[dict]):
return [sample for sample in ds if sum(sample["label"]) != 0]
if __name__ == "__main__":
# `load_dataset` might not work
# fix is here: https://github.com/huggingface/datasets/issues/7248
events_dataset = load_dataset("knowledgator/events_classification_biotech", trust_remote_code=True)
intents_data = extract_intents_data(events_dataset)
train_samples = convert_events(events_dataset["train"], intents_data)
test_samples = convert_events(events_dataset["test"], intents_data)
intents_names = [intent.name for intent in intents_data]
mask = get_low_resource_classes_mask(train_samples, intents_names)
train_samples = remove_oos(remove_low_resource_classes(train_samples, mask))
test_samples = remove_oos(remove_low_resource_classes(test_samples, mask))
events_converted = Dataset.from_dict(
{"train": train_samples, "test": test_samples, "intents": intents_data}
)
```
|