File size: 4,123 Bytes
85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d 8840aaf 85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d 85f2014 4e64e80 5b7a520 4e64e80 85f2014 c333f7d 85f2014 011ba3c 85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d 85f2014 c333f7d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 |
---
library_name: transformers
license: mit
datasets:
- allenai/peer_read
language:
- en
metrics:
- accuracy
- f1
---
# Model Card for PaperPub
*Paper pub*lication prediction based on English computer science abstracts.
## Model Details
### Model Description
PaperPub is a SciBERT ([Beltagy et al 2019](https://arxiv.org/abs/1903.10676)) model fine-tuned to predict paper acceptance from computer science abstracts.
Acceptance is modeled as a binary decision of accept or reject.
The training and evaluation data is based on the arXiv subsection of PeerRead ([Kang et al. 2018](https://aclanthology.org/N18-1149/)).
Our main use case for PaperPub is to research how attribution scores derived from acceptance predictions can inform reflecting about content and writing quality of abstracts.
- **Developed by:** Semantic Computing Group, Bielefeld University, in particular Jan-Philipp Töberg, Christoph Düsing, Jonas Belouadi and Matthias Orlikowski
- **Model type:** BERT for binary classification
- **Language(s) (NLP):** English
- **License:** MIT
- **Finetuned from model:** SciBERT
### Model Sources
We will add a public demo of PaperPub used in an application which uses attribution scores to highlight words in an abstract that contribute to acceptance/rejection predcitions.
- **Repository:** tba
- **Demo:** tba
## Uses
PaperPub can only be meaningfully used in a research setting. The model should not be used for any consequential paper quality judgements.
### Direct Use
The intended use case in research into how attribution scores computed from paper acceptance decisions reflect the abstract's content quality.
### Out-of-Scope Use
This model must not be used as part of any type of paper quality judgements, but in particular not in a peer review process. PaperPub is explicitly not meant to automate paper acceptance decisions.
## Bias, Risks, and Limitations
Bias, Risks, and Limitations are mainly related to the used datset. In addition to limitations that apply to the SciBERT pre-training corpus, our training data represents
only a very specific subset of papers. PaperPub was trained in a hackathon-like setting, so performance is not optimized and not our main goal.
### Recommendations
Users should be aware that the dataset (computer science arXive preprints from a specific period) used for fine-tuning represents a very specific idea of what papers
and in particular papers fit for publication look like.
## How to Get Started with the Model
tba
## Training Details
### Training Data
Custom stratified split of the arXiv subsection of PeerRead ([Kang et al. 2018](https://aclanthology.org/N18-1149/)). We use the data from their GitHub repository, not the Huggingface Hub version.
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Training Hyperparameters
- **Training regime:** bf16 mixed precision
- **Epochs:** 2
- **Initial Learning Rate:** 2^-5
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
Custom stratified split of the arXiv subsection of PeerRead ([Kang et al. 2018](https://aclanthology.org/N18-1149/)). We use the data from their GitHub repository, not the Huggingface Hub version.
#### Factors
Models, we compare to a naive most-frequent-class baseline.
#### Metrics
Accuracy, Macro F1
### Results
- Majority Baseline
- Acc. - 0.75
- Macro F1 - 0.43
- PaperPub
- Acc. - 0.82
- Macro F1 - 0.76
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
- **Hardware Type:** 1xA40
- **Hours used:** 0.3
- **Cloud Provider:** Private Infrastructure
- **Compute Region:** Europe
## Technical Specifications
### Compute Infrastructure
We are using an internal SLURM cluster with A40 GPUs
## Citation
tba
## Model Card Contact
[Matthias Orlikowski](https://orlikow.ski) |