JH976 commited on
Commit
920048f
·
1 Parent(s): 0725b96

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -24
README.md CHANGED
@@ -93,51 +93,34 @@ Fine-tuning is conducted on curated perovskite precursor additive datasets.
93
 
94
  <!-- This should link to a Dataset Card if possible. -->
95
 
96
- [More Information Needed]
97
 
98
- #### Factors
99
 
100
  <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
101
 
102
- [More Information Needed]
103
-
104
- #### Metrics
105
 
106
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
107
-
108
- [More Information Needed]
109
 
110
  ### Results
111
 
112
- [More Information Needed]
113
 
114
  #### Summary
115
 
 
116
 
117
 
118
- ## Model Examination [optional]
119
-
120
- <!-- Relevant interpretability work for the model goes here -->
121
-
122
- [More Information Needed]
123
-
124
  ## Environmental Impact
125
 
126
  <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
127
 
128
- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
129
-
130
- - **Hardware Type:** [More Information Needed]
131
- - **Hours used:** [More Information Needed]
132
- - **Cloud Provider:** [More Information Needed]
133
- - **Compute Region:** [More Information Needed]
134
- - **Carbon Emitted:** [More Information Needed]
135
 
136
- ## Technical Specifications [optional]
137
 
138
  ### Model Architecture and Objective
139
 
140
- [More Information Needed]
141
 
142
  ### Compute Infrastructure
143
 
 
93
 
94
  <!-- This should link to a Dataset Card if possible. -->
95
 
96
+ Performance is assessed using a benchmark dataset relevant to perovskite solar cell precursor additives
97
 
98
+ #### Factors & Metrics
99
 
100
  <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
101
 
102
+ Evaluation focuses on domain-specific factors such as material composition suggestions and additive effects on device performance. Metrics include qualitative alignment with literature findings and consistency in benchmark predictions.
 
 
103
 
 
 
 
104
 
105
  ### Results
106
 
107
+ Model performance was evaluated using a benchmark dataset related to perovskite solar cell precursor additives. Compared with other models on the same benchmark, it demonstrates superior alignment with literature findings and provides consistent, reliable suggestions for material composition and additive effects. While the results are based on this benchmark, they indicate strong potential for supporting research in perovskite solar cell chemistry.
108
 
109
  #### Summary
110
 
111
+ This model is a large language model specialized in perovskite solar cell precursor additives. It can support research by providing literature-aligned suggestions for material composition and additive effects. Evaluated on a benchmark dataset, it shows superior performance compared to other models, indicating strong potential for accelerating research and hypothesis generation. Users are encouraged to verify all outputs experimentally and consult domain experts for critical decisions.
112
 
113
 
 
 
 
 
 
 
114
  ## Environmental Impact
115
 
116
  <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
117
 
118
+ Training was conducted on a limited number of GPUs with standard energy consumption, and no significant environmental impact is expected.
 
 
 
 
 
 
119
 
 
120
 
121
  ### Model Architecture and Objective
122
 
123
+ The model is a transformer-based large language model designed for scientific text understanding and generation in the domain of perovskite solar cell precursor additives. It adopts a decoder-only architecture and is fine-tuned using instruction tuning with LoRA to provide accurate and literature-aligned suggestions for material composition, additive effects, and device performance optimization.
124
 
125
  ### Compute Infrastructure
126