File size: 7,382 Bytes
450d668
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
141ba80
450d668
141ba80
450d668
 
 
 
 
 
 
 
 
 
 
 
141ba80
 
 
 
 
 
 
 
 
 
 
 
 
450d668
 
 
141ba80
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
450d668
 
 
141ba80
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
450d668
 
 
141ba80
 
 
 
 
 
 
 
 
 
 
450d668
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7485a59
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
450d668
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
---
library_name: transformers
license: apache-2.0
base_model: allenai/specter2_base
tags:
- generated_from_trainer
metrics:
- accuracy
model-index:
- name: results
  results: []
---

<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->

# 📗 SPECTER2–MAG (Multiclass Classification on MAG Level-0 Fields of Study)

This model is a fine-tuned version of [allenai/specter2_base](https://huggingface.co/allenai/specter2_base) for multiclass bibliometric classification using MAG Fields of Study – Level 0 (SciDocs).
It achieves the following results on the evaluation set:
- Loss: 1.0598
- Accuracy: 0.8310
- Precision Micro: 0.8310
- Precision Macro: 0.8290
- Recall Micro: 0.8310
- Recall Macro: 0.8276
- F1 Micro: 0.8310
- F1 Macro: 0.8263

## Model description

This model is a fine-tuned version of SPECTER2 (`allenai/specter2_base`) adapted for multiclass classification across the 19 top-level Fields of Study (FoS) from the Microsoft Academic Graph (MAG).

The model accepts the title, abstract, or title + abstract of a scientific publication and assigns it to exactly one of the MAG Level-0 domains (e.g., Biology, Chemistry, Computer Science, Engineering, Psychology).

Key characteristics:
* Base model: allenai/specter2_base
* Task: multiclass document classification
* Labels: 19 MAG Field of Study Level-0 categories
* Activation: softmax
* Loss: CrossEntropyLoss
* Output: single best-matching FoS category

MAG Level-0 represents broad disciplinary domains designed for high-level categorization of scientific documents.

## Intended uses & limitations

### Intended uses
This multiclass MAG model is suitable for:

- Assigning publications to **top-level scientific disciplines**
- Enriching metadata in:
  - repositories  
  - research output systems  
  - funding and project datasets  
  - bibliometric dashboards  
- Supporting scientometric analyses such as:
  - broad-discipline portfolio mapping  
  - domain-level clustering  
  - modeling research diversification  
- Classifying documents when only **title/abstract** is available

The model supports inputs such as:
- **title only**
- **abstract only**
- **title + abstract** (recommended)

### Limitations
- MAG Level-0 categories are **very coarse** (e.g., *Biology*, *Medicine*, *Engineering*), and do not represent subfields.
- Documents spanning multiple fields must be forced into **one** label—an inherent limitation of multiclass classification.
- The training labels come from **MAG’s automatic field assignment pipeline**, not manual expert annotation.
- Not suitable for:
  - fine-grained subdisciplines  
  - downstream tasks requiring multilabel outputs  
  - WoS Categories or ASJC Areas (use separate models)  
  - clinical or regulatory decision-making  

Predictions should be treated as **high-level disciplinary metadata**, not detailed field classification.

## Training and evaluation data

### Source dataset: **SciDocs**

Training data comes from the **SciDocs** dataset, introduced together with the original SPECTER paper:

> **SciDocs** provides citation graphs, titles, abstracts, and **MAG Fields of Study** for scientific documents derived from MAG.  
> For this model, we use **MAG Level-0 FoS**, the 19 top-level scientific domains.

Dataset characteristics:

| Property | Value |
|---------|-------|
| Documents | ~40k scientific papers |
| Labels | 19 FoS Level-0 categories |
| Input fields | Abstract |
| Task type | Multiclass |
| Source | SciDocs (SPECTER paper) |
| License | CC-BY |

## Training procedure

### Preprocessing
- Input text constructed as:  
  `abstract`
- Tokenization using the SPECTER2 tokenizer  
- Maximum sequence length: **512 tokens**

### Model
- Base model: `allenai/specter2_base`  
- Classification head: linear layer → softmax  
- Loss: **CrossEntropyLoss**  

### Training hyperparameters

The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- num_epochs: 10

### Training results

| Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision Micro | Precision Macro | Recall Micro | Recall Macro | F1 Micro | F1 Macro |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:---------------:|:---------------:|:------------:|:------------:|:--------:|:--------:|
| 0.2603        | 1.0   | 1094 | 0.6733          | 0.8243   | 0.8243          | 0.8315          | 0.8243       | 0.8198       | 0.8243   | 0.8222   |
| 0.1779        | 2.0   | 2188 | 0.6955          | 0.8240   | 0.8240          | 0.8198          | 0.8240       | 0.8203       | 0.8240   | 0.8176   |
| 0.1628        | 3.0   | 3282 | 0.8130          | 0.8315   | 0.8315          | 0.8296          | 0.8315       | 0.8265       | 0.8315   | 0.8269   |
| 0.1136        | 4.0   | 4376 | 0.9842          | 0.8227   | 0.8227          | 0.8254          | 0.8227       | 0.8192       | 0.8227   | 0.8205   |
| 0.0666        | 5.0   | 5470 | 1.0598          | 0.8310   | 0.8310          | 0.8290          | 0.8310       | 0.8276       | 0.8310   | 0.8263   |

### Evaluation results

|                       |   precision |   recall |   f1-score |     support |
|:----------------------|------------:|---------:|-----------:|------------:|
| Art                   |    0.654867 | 0.845714 |   0.738155 |  175        |
| Biology               |    0.982222 | 0.973568 |   0.977876 |  227        |
| Business              |    0.914894 | 0.877551 |   0.895833 |  196        |
| Chemistry             |    0.97449  | 0.969543 |   0.97201  |  197        |
| Computer science      |    0.960452 | 0.894737 |   0.926431 |  190        |
| Economics             |    0.816425 | 0.782407 |   0.799054 |  216        |
| Engineering           |    0.906103 | 0.927885 |   0.916865 |  208        |
| Environmental science |    0.975369 | 0.916667 |   0.945107 |  216        |
| Geography             |    0.758454 | 0.912791 |   0.828496 |  172        |
| Geology               |    0.96729  | 0.976415 |   0.971831 |  212        |
| History               |    0.62987  | 0.518717 |   0.568915 |  187        |
| Materials science     |    0.932432 | 0.958333 |   0.945205 |  216        |
| Mathematics           |    0.938776 | 0.94359  |   0.941176 |  195        |
| Medicine              |    0.982558 | 0.923497 |   0.952113 |  183        |
| Philosophy            |    0.752874 | 0.748571 |   0.750716 |  175        |
| Physics               |    0.964824 | 0.974619 |   0.969697 |  197        |
| Political science     |    0.642512 | 0.661692 |   0.651961 |  201        |
| Psychology            |    0.806283 | 0.758621 |   0.781726 |  203        |
| Sociology             |    0.438889 | 0.427027 |   0.432877 |  185        |
| accuracy              |    0.845641 | 0.845641 |   0.845641 |    0.845641 |
| macro avg             |    0.842083 | 0.841681 |   0.840318 | 3751        |
| weighted avg          |    0.847843 | 0.845641 |   0.845311 | 3751        |


### Framework versions

- Transformers 4.57.1
- Pytorch 2.8.0+cu126
- Datasets 3.6.0
- Tokenizers 0.22.1