Datasets:

enron / enron.py
JoyCountMost's picture
Add files using upload-large-folder tool
296d8ca verified
# Copyright (c) Microsoft Corporation.
# Licensed under the MIT License.
from typing import Tuple
import datasets
import pandas as pd
from datasets import load_dataset, concatenate_datasets
from pii_leakage.arguments.ner_args import NERArgs
from pii_leakage.ner.pii_results import ListPII
from pii_leakage.ner.tagger import Tagger
from pii_leakage.ner.tagger_factory import TaggerFactory
from pii_leakage.utils.output import print_highlighted, print_dict_highlighted
from pii_leakage.utils.random import rnd_idx
from dataclasses import dataclass
@dataclass
class CustomEnronBuilder(datasets.BuilderConfig):
name: str = None
sample_duplication_rate: int = 1 # number of times a sample is repeated
shuffle_facts_seed: int = 42
pseudonymize: bool = False
class CustomEnron(datasets.GeneratorBasedBuilder):
""" A wrapper around the Enron dataset that uses anonymization. """
VERSION = datasets.Version("1.0.0")
_DESCRIPTION = "A custom wrapper for the Enron dataset."
_TEXT = "text"
BUILDER_CONFIGS = [
CustomEnronBuilder(name="undefended", sample_duplication_rate=1, version=VERSION,pseudonymize=False,
description="undefended, private data"),
CustomEnronBuilder(name="scrubbed", sample_duplication_rate=1, version=VERSION,pseudonymize=True,
description="PII replaced with anon token")
]
DEFAULT_CONFIG_NAME = "unprotected"
def __init__(self, *args, **kwargs):
self.df: pd.DataFrame = pd.DataFrame()
ner_args = NERArgs(ner='flair',
ner_model="flair/ner-english-ontonotes-large",
anon_token="<MASK>",
anonymize=kwargs.setdefault("config_name", None) == "scrubbed")
self._tagger: Tagger = TaggerFactory.from_ner_args(ner_args)
print_dict_highlighted(ner_args.__dict__)
super().__init__(*args, **kwargs)
def _info(self):
fea_dict = {self._TEXT: datasets.Value("string"),}
if self.config.pseudonymize:
fea_dict.update({entity_class: datasets.Value("string")
for entity_class in self._tagger.get_entity_classes()})
features = datasets.Features(fea_dict)
return datasets.DatasetInfo(
description=self._DESCRIPTION,
features=features
)
def _split_generators(self, dl_manager):
self.df = load_dataset("LLM-PBE/enron-email")
print("done load data")
print("self.config.pseudonymize", self.config.pseudonymize)
self.data = [item for item in self.df["train"]["text"]]
if self.config.shuffle_facts_seed > 0:
self.data = [self.data[i] for i in rnd_idx(N=len(self.data), seed=self.config.shuffle_facts_seed)]
return [
datasets.SplitGenerator( # use ~100k samples for the target model
name=datasets.Split.TRAIN,
gen_kwargs={
"split": "train",
"start": 0.0,
"end": 0.45 # default: 0.45
},
),
datasets.SplitGenerator( # use 10% of the training samples for test
name=datasets.Split.TEST,
gen_kwargs={
"split": "test",
"start": 0.45,
"end": 0.55 # default: 0.55
},
),
datasets.SplitGenerator( # Use ~110k samples for shadow models
name=datasets.Split.VALIDATION,
gen_kwargs={
"split": "validation",
"start": 0.55,
"end": 1.0 # default: 1.0
},
),
]
def _generate_examples(self, split: str, start: float, end: float):
""" Given a start and stop location, tag all PII and generate the dataset.
We use multi_gpu generation for improved speed.
"""
start_pos, end_pos = int(len(self.data) * start), int(len(self.data) * end)
print_highlighted(
f"Length of data: {len(self.data)}. Scrubbing from {start_pos} to {end_pos} (Total={end_pos - start_pos}).")
unique_identifier = start_pos
for i, text in enumerate(self.data[start_pos:end_pos]):
if self.config.pseudonymize:
pseudonymized_text, piis = self._tagger.pseudonymize(text)
# total_piis += len(piis)
# if i% 100 == 0:
# print(f"Found {total_piis} piis")
if i == 0:
print_highlighted(pseudonymized_text)
pii_annotations = {k: ListPII() for k in self._tagger.get_entity_classes()}
pii_annotations.update({k: v.dumps() for k, v in piis.group_by_class().items()})
else:
pseudonymized_text = text
pii_annotations = {}
for _ in range(self.config.sample_duplication_rate):
yield f"{unique_identifier}", {
self._TEXT: pseudonymized_text,
**pii_annotations
}
unique_identifier += 1