--- dataset_info: features: - name: id dtype: string - name: submitter dtype: string - name: authors dtype: string - name: title dtype: string - name: comments dtype: string - name: journal-ref dtype: string - name: doi dtype: string - name: report-no dtype: string - name: categories dtype: string - name: license dtype: string - name: orig_abstract dtype: string - name: versions list: - name: created dtype: string - name: version dtype: string - name: update_date dtype: string - name: authors_parsed sequence: sequence: string - name: abstract dtype: string splits: - name: train num_bytes: 295844767.0072125 num_examples: 137136 - name: test num_bytes: 63396848.15103951 num_examples: 29387 - name: validation num_bytes: 63394690.841747954 num_examples: 29386 download_size: 236976269 dataset_size: 422636305.99999994 configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* - split: validation path: data/validation-* --- # Dataset Card for hep-ph_gr-qc_primary Dataset ## Dataset Description - **Homepage:** [Kaggle arXiv Dataset Homepage](https://www.kaggle.com/Cornell-University/arxiv) - **Repository:** [hepthLlama](https://github.com/Paul-Richmond/hepthLlama) - **Paper:** [tbd](tbd) - **Point of Contact:** [Paul Richmond](mailto:p.richmond@qmul.ac.uk) ### Dataset Summary This dataset contains metadata included in arXiv submissions. ## Dataset Structure ### Languages The text in the `abstract` field of the dataset is in English, however there may be examples where the abstract also contains a translation into another language. ## Dataset Creation ### Curation Rationale The starting point was to load v193 of the Kaggle arXiv Dataset which includes arXiv submissions upto 23rd August 2024. The arXiv dataset contains the following data fields: - `id`: ArXiv ID (can be used to access the paper) - `submitter`: Who submitted the paper - `authors`: Authors of the paper - `title`: Title of the paper - `comments`: Additional info, such as number of pages and figures - `journal-ref`: Information about the journal the paper was published in - `doi`: [Digital Object Identifier](https://www.doi.org) - `report-no`: Report Number - `abstract`: The abstract of the paper - `categories`: Categories / tags in the ArXiv system To arrive at the hep-ph_gr-qc_primary dataset, the full arXiv data was filtered so that only `categories` which included 'hep-ph' or 'gr-qc' were retained. This resulted in papers that were either primarily classified as 'hep-ph' or 'gr-qc' or appeared cross-listed. For this dataset, the decision was made to focus only on papers primarily classified as either 'hep-ph' or 'gr-qc'. This meant taking only those abstracts where the first characters in `categories` were either 'hep-ph' or 'gr-qc' (see [here](https://info.arxiv.org/help/arxiv_identifier_for_services.html#indications-of-classification) for more details). We also dropped entries whose `abstract` or `comments` contained the word 'Withdrawn' or 'withdrawn' and we removed the five records which appear in the repo `LLMsForHepth/arxiv_hepth_first_overfit`. In addition, we have cleaned the data appearing in `abstract` by first replacing all occurences of '\n' with a whitespace and then removing any leading and trailing whitespace. ### Data splits The dataset is split into a training, validation and test set with split percentages 70%, 15% and 15%. This was done by applying `train_test_split` twice (both with `seed=42`). The final split sizes are as follows: | Train | Test | Validation | |:---:|:---:|:---:| |137,136 | 29,387| 29,386 |