Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,90 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
tags:
|
| 3 |
+
- text-classification
|
| 4 |
+
- roberta
|
| 5 |
+
- scientific-abstracts
|
| 6 |
+
- multi-class
|
| 7 |
+
- research-field-classification
|
| 8 |
+
datasets:
|
| 9 |
+
- ScientificArticleAbstract_Classification
|
| 10 |
+
license: apache-2.0
|
| 11 |
+
model-index:
|
| 12 |
+
- name: ScientificTextClassification_ResearchField
|
| 13 |
+
results:
|
| 14 |
+
- task:
|
| 15 |
+
name: Text Classification
|
| 16 |
+
type: text-classification
|
| 17 |
+
metrics:
|
| 18 |
+
- type: accuracy
|
| 19 |
+
value: 0.941
|
| 20 |
+
name: Accuracy (Top-1)
|
| 21 |
+
- type: macro_f1
|
| 22 |
+
value: 0.935
|
| 23 |
+
name: Macro F1 Score
|
| 24 |
+
---
|
| 25 |
+
|
| 26 |
+
# ScientificTextClassification_ResearchField
|
| 27 |
+
|
| 28 |
+
## 📚 Overview
|
| 29 |
+
|
| 30 |
+
This is a **RoBERTa-base** model fine-tuned for the complex task of multi-class classification of scientific article abstracts. The model predicts the **primary research field** (e.g., Physics, Biology, Computer Science) based solely on the abstract text, serving as a powerful tool for automated journal indexing and literature review organization.
|
| 31 |
+
|
| 32 |
+
## 🧠 Model Architecture
|
| 33 |
+
|
| 34 |
+
The choice of RoBERTa ensures enhanced robustness and better handling of long-range dependencies common in technical and scientific prose.
|
| 35 |
+
|
| 36 |
+
* **Base Model:** `roberta-base` (an optimized BERT approach without the next-sentence prediction objective).
|
| 37 |
+
* **Classification Head:** Outputs 8 distinct categories (`num_labels: 8`).
|
| 38 |
+
* **Input Data:** Detailed scientific abstracts from diverse journals.
|
| 39 |
+
* **Output:** A probability distribution over the 8 classes: Physics, Chemistry, Medicine, Computer Science, Biology, Geoscience, Materials Science, and Engineering.
|
| 40 |
+
* **Training Dataset:** **ScientificArticleAbstract_Classification**, providing abstracts linked to their high-level research disciplines.
|
| 41 |
+
|
| 42 |
+
## 🎯 Intended Use
|
| 43 |
+
|
| 44 |
+
The model offers utility in several scientific and information retrieval contexts:
|
| 45 |
+
|
| 46 |
+
1. **Automated Library and Repository Indexing:** Rapidly and accurately tagging new publications with their correct discipline.
|
| 47 |
+
2. **Literature Review Automation:** Filtering large databases of articles to focus on specific fields.
|
| 48 |
+
3. **Grant Proposal Routing:** Assisting research institutions in routing incoming proposals to the appropriate review panel or expert based on the summary.
|
| 49 |
+
4. **Trend Analysis:** Tracking the volume and convergence of research across different fields.
|
| 50 |
+
|
| 51 |
+
## ⚠️ Limitations
|
| 52 |
+
|
| 53 |
+
1. **Interdisciplinary Papers:** The model performs single-label classification. It may struggle with highly interdisciplinary abstracts that bridge two or more distinct fields (e.g., computational chemistry or bio-engineering).
|
| 54 |
+
2. **Vocabulary Drift:** Scientific terminology evolves quickly. New sub-disciplines or extremely novel concepts may not be classified correctly until the model is retrained.
|
| 55 |
+
3. **Class Imbalance:** If the underlying distribution of the eight fields in the real world shifts significantly from the training set, performance may vary.
|
| 56 |
+
|
| 57 |
+
### MODEL 3: **EcommerceAspectSentiment_BART**
|
| 58 |
+
|
| 59 |
+
This model is a BART-large sequence-to-sequence model fine-tuned for abstractive multi-aspect sentiment summarization based on Dataset 3 (EcommerceCustomerReview\_MultiAspectRating).
|
| 60 |
+
|
| 61 |
+
#### config.json
|
| 62 |
+
|
| 63 |
+
```json
|
| 64 |
+
{
|
| 65 |
+
"_name_or_path": "facebook/bart-large",
|
| 66 |
+
"architectures": [
|
| 67 |
+
"BartForConditionalGeneration"
|
| 68 |
+
],
|
| 69 |
+
"model_type": "bart",
|
| 70 |
+
"vocab_size": 50265,
|
| 71 |
+
"d_model": 1024,
|
| 72 |
+
"encoder_layers": 12,
|
| 73 |
+
"decoder_layers": 12,
|
| 74 |
+
"encoder_attention_heads": 16,
|
| 75 |
+
"decoder_attention_heads": 16,
|
| 76 |
+
"encoder_ffn_dim": 4096,
|
| 77 |
+
"decoder_ffn_dim": 4096,
|
| 78 |
+
"dropout": 0.1,
|
| 79 |
+
"activation_function": "gelu",
|
| 80 |
+
"init_std": 0.02,
|
| 81 |
+
"num_labels": 3,
|
| 82 |
+
"max_position_embeddings": 1024,
|
| 83 |
+
"eos_token_id": 2,
|
| 84 |
+
"bos_token_id": 0,
|
| 85 |
+
"pad_token_id": 1,
|
| 86 |
+
"is_encoder_decoder": true,
|
| 87 |
+
"scale_embedding": false,
|
| 88 |
+
"forced_eos_token_id": 2,
|
| 89 |
+
"transformers_version": "4.35.2"
|
| 90 |
+
}
|