Tasfiya025 commited on
Commit
0ab7dc9
·
verified ·
1 Parent(s): 05c982c

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +90 -0
README.md ADDED
@@ -0,0 +1,90 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - text-classification
4
+ - roberta
5
+ - scientific-abstracts
6
+ - multi-class
7
+ - research-field-classification
8
+ datasets:
9
+ - ScientificArticleAbstract_Classification
10
+ license: apache-2.0
11
+ model-index:
12
+ - name: ScientificTextClassification_ResearchField
13
+ results:
14
+ - task:
15
+ name: Text Classification
16
+ type: text-classification
17
+ metrics:
18
+ - type: accuracy
19
+ value: 0.941
20
+ name: Accuracy (Top-1)
21
+ - type: macro_f1
22
+ value: 0.935
23
+ name: Macro F1 Score
24
+ ---
25
+
26
+ # ScientificTextClassification_ResearchField
27
+
28
+ ## 📚 Overview
29
+
30
+ This is a **RoBERTa-base** model fine-tuned for the complex task of multi-class classification of scientific article abstracts. The model predicts the **primary research field** (e.g., Physics, Biology, Computer Science) based solely on the abstract text, serving as a powerful tool for automated journal indexing and literature review organization.
31
+
32
+ ## 🧠 Model Architecture
33
+
34
+ The choice of RoBERTa ensures enhanced robustness and better handling of long-range dependencies common in technical and scientific prose.
35
+
36
+ * **Base Model:** `roberta-base` (an optimized BERT approach without the next-sentence prediction objective).
37
+ * **Classification Head:** Outputs 8 distinct categories (`num_labels: 8`).
38
+ * **Input Data:** Detailed scientific abstracts from diverse journals.
39
+ * **Output:** A probability distribution over the 8 classes: Physics, Chemistry, Medicine, Computer Science, Biology, Geoscience, Materials Science, and Engineering.
40
+ * **Training Dataset:** **ScientificArticleAbstract_Classification**, providing abstracts linked to their high-level research disciplines.
41
+
42
+ ## 🎯 Intended Use
43
+
44
+ The model offers utility in several scientific and information retrieval contexts:
45
+
46
+ 1. **Automated Library and Repository Indexing:** Rapidly and accurately tagging new publications with their correct discipline.
47
+ 2. **Literature Review Automation:** Filtering large databases of articles to focus on specific fields.
48
+ 3. **Grant Proposal Routing:** Assisting research institutions in routing incoming proposals to the appropriate review panel or expert based on the summary.
49
+ 4. **Trend Analysis:** Tracking the volume and convergence of research across different fields.
50
+
51
+ ## ⚠️ Limitations
52
+
53
+ 1. **Interdisciplinary Papers:** The model performs single-label classification. It may struggle with highly interdisciplinary abstracts that bridge two or more distinct fields (e.g., computational chemistry or bio-engineering).
54
+ 2. **Vocabulary Drift:** Scientific terminology evolves quickly. New sub-disciplines or extremely novel concepts may not be classified correctly until the model is retrained.
55
+ 3. **Class Imbalance:** If the underlying distribution of the eight fields in the real world shifts significantly from the training set, performance may vary.
56
+
57
+ ### MODEL 3: **EcommerceAspectSentiment_BART**
58
+
59
+ This model is a BART-large sequence-to-sequence model fine-tuned for abstractive multi-aspect sentiment summarization based on Dataset 3 (EcommerceCustomerReview\_MultiAspectRating).
60
+
61
+ #### config.json
62
+
63
+ ```json
64
+ {
65
+ "_name_or_path": "facebook/bart-large",
66
+ "architectures": [
67
+ "BartForConditionalGeneration"
68
+ ],
69
+ "model_type": "bart",
70
+ "vocab_size": 50265,
71
+ "d_model": 1024,
72
+ "encoder_layers": 12,
73
+ "decoder_layers": 12,
74
+ "encoder_attention_heads": 16,
75
+ "decoder_attention_heads": 16,
76
+ "encoder_ffn_dim": 4096,
77
+ "decoder_ffn_dim": 4096,
78
+ "dropout": 0.1,
79
+ "activation_function": "gelu",
80
+ "init_std": 0.02,
81
+ "num_labels": 3,
82
+ "max_position_embeddings": 1024,
83
+ "eos_token_id": 2,
84
+ "bos_token_id": 0,
85
+ "pad_token_id": 1,
86
+ "is_encoder_decoder": true,
87
+ "scale_embedding": false,
88
+ "forced_eos_token_id": 2,
89
+ "transformers_version": "4.35.2"
90
+ }