JH976 commited on
Commit
1a4aa35
·
1 Parent(s): 920048f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -29
README.md CHANGED
@@ -36,14 +36,14 @@ This model is a domain-specific large language model fine-tuned from QwQ-32B, sp
36
 
37
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
38
 
39
- You can run it based on LLaMA-Factory or other frameworks that can call LLMs.
40
 
41
 
42
  ### Out-of-Scope Use
43
 
44
  <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
45
 
46
- This model is not suitable for fields other than perovskite solar cells.
47
 
48
  ## Bias, Risks, and Limitations
49
 
@@ -124,42 +124,25 @@ The model is a transformer-based large language model designed for scientific te
124
 
125
  ### Compute Infrastructure
126
 
127
- [More Information Needed]
128
 
129
  #### Hardware
130
 
131
- [More Information Needed]
132
 
133
  #### Software
134
 
135
- [More Information Needed]
136
 
137
- ## Citation [optional]
138
 
139
  <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
140
 
141
  **BibTeX:**
142
 
143
- [More Information Needed]
144
-
145
- **APA:**
146
-
147
- [More Information Needed]
148
-
149
- ## Glossary [optional]
150
-
151
- <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
152
-
153
- [More Information Needed]
154
-
155
- ## More Information [optional]
156
-
157
- [More Information Needed]
158
-
159
- ## Model Card Authors [optional]
160
-
161
- [More Information Needed]
162
-
163
- ## Model Card Contact
164
-
165
- [More Information Needed]
 
36
 
37
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
38
 
39
+ The model can be run using LLaMA-Factory or any other framework capable of calling large language models. Users are recommended to follow the framework’s instructions for loading and running the model, ensuring sufficient GPU memory and compatibility with LoRA-adapted weights.
40
 
41
 
42
  ### Out-of-Scope Use
43
 
44
  <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
45
 
46
+ This model is specifically designed for perovskite solar cell precursor additives and is not intended for use in other fields. Using it outside this domain may produce unreliable or inaccurate suggestions, and could lead to incorrect conclusions if applied to unrelated materials or chemical systems.
47
 
48
  ## Bias, Risks, and Limitations
49
 
 
124
 
125
  ### Compute Infrastructure
126
 
127
+ Training was conducted on high-performance GPUs with bfloat16 precision, utilizing gradient accumulation and FlashAttention2 for efficiency. Users can run the model on GPUs with sufficient memory to accommodate LoRA-adapted weights.
128
 
129
  #### Hardware
130
 
131
+ Training is performed on high-performance NVIDIA GPUs (e.g., H100 or equivalent). For inference, users should use GPUs with sufficient memory to load LoRA-adapted weights.
132
 
133
  #### Software
134
 
135
+ The model was trained and can be run using Python (>=3.10), PyTorch, and LLaMA-Factory. Additional libraries such as Transformers and FlashAttention2 are used for optimized performance.
136
 
137
+ ## Citation
138
 
139
  <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
140
 
141
  **BibTeX:**
142
 
143
+ @article{wang2025perovskite,
144
+ title={Perovskite-R1: A Domain-Specialized LLM for Intelligent Discovery of Precursor Additives and Experimental Design},
145
+ author={Wang, Xin-De and Chen, Zhi-Rui and Guo, Peng-Jie and Gao, Ze-Feng and Mu, Cheng and Lu, Zhong-Yi},
146
+ journal={arXiv preprint arXiv:2507.16307},
147
+ year={2025}
148
+ }