Text Classification
Transformers
Safetensors
English
bert
File size: 4,123 Bytes
85f2014
 
c333f7d
 
 
 
 
 
 
 
85f2014
 
c333f7d
85f2014
c333f7d
85f2014
 
 
 
 
c333f7d
 
 
 
85f2014
c333f7d
 
 
 
 
85f2014
c333f7d
85f2014
c333f7d
85f2014
c333f7d
 
85f2014
 
 
c333f7d
85f2014
 
 
c333f7d
85f2014
 
 
c333f7d
85f2014
 
 
c333f7d
8840aaf
85f2014
 
 
c333f7d
 
85f2014
 
 
c333f7d
85f2014
 
 
 
 
c333f7d
85f2014
 
 
 
 
 
 
 
c333f7d
 
 
 
85f2014
 
 
 
 
 
 
 
c333f7d
85f2014
 
 
c333f7d
85f2014
 
 
c333f7d
85f2014
 
 
4e64e80
 
 
5b7a520
4e64e80
 
85f2014
 
 
 
 
 
c333f7d
 
 
 
85f2014
011ba3c
85f2014
 
 
c333f7d
85f2014
c333f7d
85f2014
c333f7d
85f2014
 
 
c333f7d
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
---
library_name: transformers
license: mit
datasets:
- allenai/peer_read
language:
- en
metrics:
- accuracy
- f1
---

# Model Card for PaperPub

*Paper pub*lication prediction based on English computer science abstracts.

## Model Details

### Model Description

PaperPub is a SciBERT ([Beltagy et al 2019](https://arxiv.org/abs/1903.10676)) model fine-tuned to predict paper acceptance from computer science abstracts. 
Acceptance is modeled as a binary decision of accept or reject.
The training and evaluation data is based on the arXiv subsection of PeerRead ([Kang et al. 2018](https://aclanthology.org/N18-1149/)). 
Our main use case for PaperPub is to research how attribution scores derived from acceptance predictions can inform reflecting about content and writing quality of abstracts. 

- **Developed by:** Semantic Computing Group, Bielefeld University, in particular Jan-Philipp Töberg, Christoph Düsing, Jonas Belouadi and Matthias Orlikowski
- **Model type:** BERT for binary classification
- **Language(s) (NLP):** English
- **License:** MIT
- **Finetuned from model:** SciBERT

### Model Sources

We will add a public demo of PaperPub used in an application which uses attribution scores to highlight words in an abstract that contribute to acceptance/rejection predcitions.

- **Repository:** tba
- **Demo:** tba

## Uses

PaperPub can only be meaningfully used in a research setting. The model should not be used for any consequential paper quality judgements.

### Direct Use

The intended use case in research into how attribution scores computed from paper acceptance decisions reflect the abstract's content quality.

### Out-of-Scope Use

This model must not be used as part of any type of paper quality judgements, but in particular not in a peer review process. PaperPub is explicitly not meant to automate paper acceptance decisions.

## Bias, Risks, and Limitations

Bias, Risks, and Limitations are mainly related to the used datset. In addition to limitations that apply to the SciBERT pre-training corpus, our training data represents
only a very specific subset of papers. PaperPub was trained in a hackathon-like setting, so performance is not optimized and not our main goal.

### Recommendations

Users should be aware that the dataset (computer science arXive preprints from a specific period) used for fine-tuning represents a very specific idea of what papers 
and in particular papers fit for publication look like.

## How to Get Started with the Model

tba

## Training Details

### Training Data

Custom stratified split of the arXiv subsection of PeerRead ([Kang et al. 2018](https://aclanthology.org/N18-1149/)). We use the data from their GitHub repository, not the Huggingface Hub version.

### Training Procedure

<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->


#### Training Hyperparameters

- **Training regime:** bf16 mixed precision
- **Epochs:** 2
- **Initial Learning Rate:** 2^-5
 
## Evaluation

<!-- This section describes the evaluation protocols and provides the results. -->

### Testing Data, Factors & Metrics

#### Testing Data

Custom stratified split of the arXiv subsection of PeerRead ([Kang et al. 2018](https://aclanthology.org/N18-1149/)). We use the data from their GitHub repository, not the Huggingface Hub version.

#### Factors

Models, we compare to a naive most-frequent-class baseline.

#### Metrics

Accuracy, Macro F1

### Results

- Majority Baseline
  - Acc. - 0.75
  - Macro F1 - 0.43
- PaperPub
  - Acc. - 0.82
  - Macro F1 - 0.76


## Environmental Impact

<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->

- **Hardware Type:** 1xA40
- **Hours used:** 0.3
- **Cloud Provider:** Private Infrastructure
- **Compute Region:** Europe

## Technical Specifications

### Compute Infrastructure

We are using an internal SLURM cluster with A40 GPUs

## Citation

tba

## Model Card Contact

[Matthias Orlikowski](https://orlikow.ski)