| | """ |
| | The Harvard USPTO Patent Dataset (HUPD) is a large-scale, well-structured, and multi-purpose corpus |
| | of English-language patent applications filed to the United States Patent and Trademark Office (USPTO) |
| | between 2004 and 2018. With more than 4.5 million patent documents, HUPD is two to three times larger |
| | than comparable corpora. Unlike other NLP patent datasets, HUPD contains the inventor-submitted versions |
| | of patent applications, not the final versions of granted patents, allowing us to study patentability at |
| | the time of filing using NLP methods for the first time. |
| | """ |
| |
|
| | from __future__ import absolute_import, division, print_function |
| |
|
| | import os |
| | import datetime |
| | import pandas as pd |
| | import numpy as np |
| | from pathlib import Path |
| | try: |
| | import ujson as json |
| | except: |
| | import json |
| |
|
| | import datasets |
| |
|
| |
|
| | _CITATION = """\ |
| | @InProceedings{suzgun2021:hupd, |
| | title = {The Harvard USPTO Patent Dataset}, |
| | authors={Mirac Suzgun and Suproteem Sarkar and Luke Melas-Kyriazi and Scott Kominers and Stuart Shieber}, |
| | year={2021} |
| | } |
| | """ |
| |
|
| | _DESCRIPTION = """ |
| | The Harvard USPTO Patent Dataset (HUPD) is a large-scale, well-structured, and multi-purpose corpus |
| | of English-language patent applications filed to the United States Patent and Trademark Office (USPTO) |
| | between 2004 and 2018. With more than 4.5 million patent documents, HUPD is two to three times larger |
| | than comparable corpora. Unlike other NLP patent datasets, HUPD contains the inventor-submitted versions |
| | of patent applications, not the final versions of granted patents, allowing us to study patentability at |
| | the time of filing using NLP methods for the first time. |
| | """ |
| |
|
| | RANDOM_STATE = 1729 |
| |
|
| | _FEATURES = [ |
| | "patent_number", |
| | "decision", |
| | "title", |
| | "abstract", |
| | "claims", |
| | "background", |
| | "summary", |
| | "description", |
| | "cpc_label", |
| | "ipc_label", |
| | "filing_date", |
| | "patent_issue_date", |
| | "date_published", |
| | "examiner_id" |
| | ] |
| |
|
| |
|
| | def str_to_date(s): |
| | """A helper function to convert strings to dates""" |
| | return datetime.datetime.strptime(s, '%Y-%m-%d') |
| |
|
| |
|
| | class PatentsConfig(datasets.BuilderConfig): |
| | """BuilderConfig for Patents""" |
| |
|
| | def __init__( |
| | self, |
| | metadata_url: str, |
| | data_url: str, |
| | data_dir: str, |
| | ipcr_label: str = None, |
| | cpc_label: str = None, |
| | train_filing_start_date: str = None, |
| | train_filing_end_date: str = None, |
| | val_filing_start_date: str = None, |
| | val_filing_end_date: str = None, |
| | query_string: str = None, |
| | val_set_balancer=False, |
| | uniform_split=False, |
| | force_extract=False, |
| | **kwargs |
| | ): |
| | """ |
| | If train_filing_end_date is None, then a random train-val split will be used. If it is |
| | specified, then the specified date range will be used for the split. If train_filing_end_date |
| | if specified and val_filing_start_date is not specifed, then val_filing_start_date defaults to |
| | train_filing_end_date. |
| | |
| | Args: |
| | metadata_url: `string`, url from which to download the metadata file |
| | data_url: `string`, url from which to download the json files |
| | data_dir: `string`, folder (in cache) in which downloaded json files are stored |
| | ipcr_label: International Patent Classification code |
| | cpc_label: Cooperative Patent Classification code |
| | train_filing_start_date: Start date for patents in train set (and val set if random split is used) |
| | train_filing_end_date: End date for patents in train set |
| | val_filing_start_date: Start date for patents in val set |
| | val_filing_end_date: End date for patents in val set (and train set if random split is used) |
| | force_extract: Extract only the relevant years if this parameter is used. |
| | **kwargs: keyword arguments forwarded to super |
| | """ |
| | super().__init__(**kwargs) |
| | self.metadata_url = metadata_url |
| | self.data_url = data_url |
| | self.data_dir = data_dir |
| | self.ipcr_label = ipcr_label |
| | self.cpc_label = cpc_label |
| | self.train_filing_start_date = train_filing_start_date |
| | self.train_filing_end_date = train_filing_end_date |
| | self.val_filing_start_date = val_filing_start_date |
| | self.val_filing_end_date = val_filing_end_date |
| | self.query_string = query_string |
| | self.val_set_balancer = val_set_balancer |
| | self.uniform_split = uniform_split |
| | self.force_extract = force_extract |
| |
|
| |
|
| | class Patents(datasets.GeneratorBasedBuilder): |
| | _DESCRIPTION |
| |
|
| | VERSION = datasets.Version("1.0.2") |
| |
|
| | |
| | |
| | |
| | BUILDER_CONFIG_CLASS = PatentsConfig |
| | BUILDER_CONFIGS = [ |
| | PatentsConfig( |
| | name="sample", |
| | description="Patent data from January 2016, for debugging", |
| | metadata_url="https://huggingface.co/datasets/HUPD/hupd/resolve/main/hupd_metadata_jan16_2022-02-22.feather", |
| | data_url="https://huggingface.co/datasets/HUPD/hupd/resolve/main/data/sample-jan-2016.tar.gz", |
| | data_dir="sample", |
| | ), |
| | PatentsConfig( |
| | name="all", |
| | description="Patent data from all years (2004-2018)", |
| | metadata_url="https://huggingface.co/datasets/HUPD/hupd/resolve/main/hupd_metadata_2022-02-22.feather", |
| | data_url="https://huggingface.co/datasets/HUPD/hupd/resolve/main/data/all-years.tar", |
| | data_dir="data", |
| | ), |
| | ] |
| |
|
| | def _info(self): |
| | return datasets.DatasetInfo( |
| | |
| | description=_DESCRIPTION, |
| | |
| | features=datasets.Features( |
| | {k: datasets.Value("string") for k in _FEATURES} |
| | ), |
| | |
| | |
| | |
| | supervised_keys=("claims", "decision"), |
| | homepage="https://github.com/suzgunmirac/hupd", |
| | citation=_CITATION, |
| | ) |
| |
|
| | def _split_generators(self, dl_manager: datasets.DownloadManager): |
| | """Returns SplitGenerators.""" |
| | print(f'Loading dataset with config: {self.config}') |
| |
|
| | |
| | |
| | metadata_url = self.config.metadata_url |
| | metadata_file = dl_manager.download_and_extract(self.config.metadata_url) |
| | print(f'Using metadata file: {metadata_file}') |
| |
|
| | |
| | |
| | |
| | download_dir = dl_manager.download_and_extract(self.config.data_url) |
| | json_dir = os.path.join(download_dir, self.config.data_dir) |
| |
|
| | |
| | print(f'Reading metadata file: {metadata_file}') |
| | if metadata_url.endswith('.feather'): |
| | df = pd.read_feather(metadata_file) |
| | elif metadata_url.endswith('.csv'): |
| | df = pd.read_csv(metadata_file) |
| | elif metadata_url.endswith('.tsv'): |
| | df = pd.read_csv(metadata_file, delimiter='\t') |
| | elif metadata_url.endswith('.pickle'): |
| | df = pd.read_pickle(metadata_file) |
| | else: |
| | raise ValueError(f'Metadata file invalid: {metadata_url}') |
| |
|
| | |
| | if self.config.ipcr_label: |
| | print(f'Filtering by IPCR label: {self.config.ipcr_label}') |
| | df = df[df['main_ipcr_label'].str.startswith(self.config.ipcr_label)] |
| | elif self.config.cpc_label: |
| | print(f'Filtering by CPC label: {self.config.cpc_label}') |
| | df = df[df['main_cpc_label'].str.startswith(self.config.cpc_label)] |
| |
|
| | |
| | if self.config.query_string: |
| | df = df.query(self.config.query_string) |
| | |
| | if self.config.force_extract: |
| | if self.config.name == 'all': |
| | if self.config.train_filing_start_date and self.config.val_filing_end_date: |
| | if self.config.train_filing_end_date and self.config.val_filing_start_date: |
| | training_year_range = set(range(int(self.config.train_filing_start_date[:4]), int(self.config.train_filing_end_date[:4]) + 1)) |
| | validation_year_range = set(range(int(self.config.val_filing_start_date[:4]), int(self.config.val_filing_end_date[:4]) + 1)) |
| | full_year_range = training_year_range.union(validation_year_range) |
| | else: |
| | full_year_range = set(range(int(self.config.train_filing_start_date[:4]), int(self.config.val_filing_end_date[:4]) + 1)) |
| | else: |
| | full_year_range = set(range(2004, 2019)) |
| | |
| |
|
| | import tarfile |
| | for year in full_year_range: |
| | tar_file_path = f'{json_dir}/{year}.tar.gz' |
| | print(f'Extracting {tar_file_path}') |
| | |
| | tar_file = tarfile.open(tar_file_path) |
| | |
| | tar_file.extractall(f'{json_dir}') |
| | tar_file.close() |
| |
|
| | |
| | if self.config.uniform_split: |
| |
|
| | |
| | if self.config.train_filing_start_date: |
| | df = df[df['filing_date'] >= self.config.train_filing_start_date] |
| | if self.config.val_filing_end_date: |
| | df = df[df['filing_date'] <= self.config.val_filing_end_date] |
| | df = df.sample(frac=1.0, random_state=RANDOM_STATE) |
| | num_train_samples = int(len(df) * 0.85) |
| | train_df = df.iloc[0:num_train_samples] |
| | val_df = df.iloc[num_train_samples:-1] |
| |
|
| | else: |
| |
|
| | |
| | if not (self.config.train_filing_start_date and self.config.train_filing_end_date and |
| | self.config.val_filing_start_date and self.config.train_filing_end_date): |
| | raise ValueError("Please either use uniform_split or specify your exact \ |
| | training and validation split dates.") |
| |
|
| | |
| | print(f'Filtering train dataset by filing start date: {self.config.train_filing_start_date}') |
| | print(f'Filtering train dataset by filing end date: {self.config.train_filing_end_date}') |
| | print(f'Filtering val dataset by filing start date: {self.config.val_filing_start_date}') |
| | print(f'Filtering val dataset by filing end date: {self.config.val_filing_end_date}') |
| | train_df = df[ |
| | (df['filing_date'] >= self.config.train_filing_start_date) & |
| | (df['filing_date'] < self.config.train_filing_end_date) |
| | ] |
| | val_df = df[ |
| | (df['filing_date'] >= self.config.val_filing_start_date) & |
| | (df['filing_date'] < self.config.val_filing_end_date) |
| | ] |
| |
|
| | |
| | if self.config.val_set_balancer: |
| | rejected_df = val_df[val_df.status == 'REJECTED'] |
| | num_rejected = len(rejected_df) |
| | accepted_df = val_df[val_df.status == 'ACCEPTED'] |
| | num_accepted = len(accepted_df) |
| | if num_rejected < num_accepted: |
| | accepted_df = accepted_df.sample(frac=1.0, random_state=RANDOM_STATE) |
| | accepted_df = accepted_df[:num_rejected] |
| | else: |
| | rejected_df = rejected_df.sample(frac=1.0, random_state=RANDOM_STATE) |
| | rejected_df = rejected_df[:num_accepted] |
| | val_df = pd.concat([rejected_df, accepted_df]) |
| |
|
| | return [ |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TRAIN, |
| | gen_kwargs=dict( |
| | df=train_df, |
| | json_dir=json_dir, |
| | split='train', |
| | ), |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split.VALIDATION, |
| | gen_kwargs=dict( |
| | df=val_df, |
| | json_dir=json_dir, |
| | split='val', |
| | ), |
| | ), |
| | ] |
| |
|
| | def _generate_examples(self, df, json_dir, split): |
| | """ Yields examples by loading JSON files containing patent applications. """ |
| |
|
| | |
| | for id_, x in enumerate(df.itertuples()): |
| |
|
| | |
| | application_year = str(x.filing_date.year) |
| | application_number = x.application_number |
| | filepath = os.path.join(json_dir, application_year, application_number + '.json') |
| | try: |
| | with open(filepath, 'r') as f: |
| | patent = json.load(f) |
| | except Exception as e: |
| | print('------------') |
| | print(f'ERROR WITH {filepath}\n') |
| | print(repr(e)) |
| | print() |
| | yield id_, {k: "error" for k in _FEATURES} |
| |
|
| | |
| | decision = x.decision |
| | yield id_, { |
| | "patent_number": application_number, |
| | "decision": patent["decision"], |
| | "title": patent["title"], |
| | "abstract": patent["abstract"], |
| | "claims": patent["claims"], |
| | "description": patent["full_description"], |
| | "background": patent["background"], |
| | "summary": patent["summary"], |
| | "cpc_label": patent["main_cpc_label"], |
| | 'filing_date': patent['filing_date'], |
| | 'patent_issue_date': patent['patent_issue_date'], |
| | 'date_published': patent['date_published'], |
| | 'examiner_id': patent['examiner_id'], |
| | "ipc_label": patent["main_ipcr_label"], |
| | } |
| |
|