File size: 4,727 Bytes
db04e21
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
dc9ab90
 
 
 
 
 
 
9efb4ad
dc9ab90
9efb4ad
dc9ab90
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
---
dataset_info:
  features:
  - name: id
    dtype: string
  - name: submitter
    dtype: string
  - name: authors
    dtype: string
  - name: title
    dtype: string
  - name: comments
    dtype: string
  - name: journal-ref
    dtype: string
  - name: doi
    dtype: string
  - name: report-no
    dtype: string
  - name: categories
    dtype: string
  - name: license
    dtype: string
  - name: orig_abstract
    dtype: string
  - name: versions
    list:
    - name: created
      dtype: string
    - name: version
      dtype: string
  - name: update_date
    dtype: string
  - name: authors_parsed
    sequence:
      sequence: string
  - name: abstract
    dtype: string
  splits:
  - name: train
    num_bytes: 147667993.3685569
    num_examples: 73768
  - name: test
    num_bytes: 31644285.315721553
    num_examples: 15808
  - name: validation
    num_bytes: 31644285.315721553
    num_examples: 15808
  download_size: 115280347
  dataset_size: 210956563.99999997
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
  - split: test
    path: data/test-*
  - split: validation
    path: data/validation-*
---


# Dataset Card for arxiv_hep-th_primary Dataset

## Dataset Description

- **Homepage:** [Kaggle arXiv Dataset Homepage](https://www.kaggle.com/Cornell-University/arxiv)
- **Repository:** [FeynTune](https://github.com/Paul-Richmond/FeynTune)
- **Paper:** [tbd](tbd)
<!-- - **Point of Contact:** [Paul Richmond](mailto:p.richmond@qmul.ac.uk) -->

### Dataset Summary
This dataset contains metadata included in arXiv submissions.

## Dataset Structure

An example from the dataset looks as follows:

```
{'id': '0908.2896',
 'submitter': 'Paul Richmond',
 'authors': 'Neil Lambert, Paul Richmond',
 'title': 'M2-Branes and Background Fields',
 'comments': '19 pages',
 'journal-ref': 'JHEP 0910:084,2009',
 'doi': '10.1088/1126-6708/2009/10/084',
 'report-no': None,
 'categories': 'hep-th',
 'license': 'http://arxiv.org/licenses/nonexclusive-distrib/1.0/',
 'abstract': '  We discuss the coupling of multiple M2-branes to the background 3-form and\n6-form gauge fields of eleven-dimensional supergravity, including the coupling\nof the Fermions. In particular we show in detail how a natural generalization\nof the Myers flux-terms, along with the resulting curvature of the background\nmetric, leads to mass terms in the effective field theory.\n',
 'versions': [{'created': 'Thu, 20 Aug 2009 14:23:37 GMT', 'version': 'v1'}],
 'update_date': '2009-11-09',
 'authors_parsed': [['Lambert', 'Neil', ''], ['Richmond', 'Paul', '']]}
```

### Languages

The text in the `abstract` field of the dataset is in English, however there may be examples 
where the abstract also contains a translation into another language. 

## Dataset Creation

### Curation Rationale
The starting point was to load v193 of the Kaggle arXiv Dataset which includes arXiv submissions upto 23rd August 2024.
The arXiv dataset contains the following data fields:
- `id`: ArXiv ID (can be used to access the paper)
- `submitter`: Who submitted the paper
- `authors`: Authors of the paper
- `title`: Title of the paper
- `comments`: Additional info, such as number of pages and figures
- `journal-ref`: Information about the journal the paper was published in
- `doi`: [Digital Object Identifier](https://www.doi.org)
- `report-no`: Report Number
- `abstract`: The abstract of the paper
- `categories`: Categories / tags in the ArXiv system

To arrive at the arxiv_hep-th_primary dataset, the full arXiv data 
was filtered so that only `categories` which included 'hep-th' were retained.
This resulted in papers that were either primarily classified as 'hep-th' or appeared cross-listed.
For this dataset, the decision was made to focus only on papers primarily classified as 'hep-th'. 
This meant taking only those abstracts where the first characters in `categories` were 'hep-th' 
(see [here](https://info.arxiv.org/help/arxiv_identifier_for_services.html#indications-of-classification) for more details).

We also dropped entries whose `abstract` or `comments` contained the word 'Withdrawn' or 'withdrawn' and we removed the five records which appear in the repo `LLMsForHepth/arxiv_hepth_first_overfit`.

In addition, we have cleaned the data appearing in `abstract` by first replacing all occurences of '\n' with a whitespace and then removing any leading and trailing whitespace.

### Data splits

The dataset is split into a training, validation and test set with split percentages 70%, 15% and 15%. This was done by applying `train_test_split` twice (both with `seed=42`).
The final split sizes are as follows:

| Train | Test | Validation |
|:---:|:---:|:---:|
|73,768 | 15,808| 15,808   |