Update README.md
Browse files
README.md
CHANGED
|
@@ -36,14 +36,14 @@ This model is a domain-specific large language model fine-tuned from QwQ-32B, sp
|
|
| 36 |
|
| 37 |
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
| 38 |
|
| 39 |
-
|
| 40 |
|
| 41 |
|
| 42 |
### Out-of-Scope Use
|
| 43 |
|
| 44 |
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
|
| 45 |
|
| 46 |
-
This model is not
|
| 47 |
|
| 48 |
## Bias, Risks, and Limitations
|
| 49 |
|
|
@@ -124,42 +124,25 @@ The model is a transformer-based large language model designed for scientific te
|
|
| 124 |
|
| 125 |
### Compute Infrastructure
|
| 126 |
|
| 127 |
-
|
| 128 |
|
| 129 |
#### Hardware
|
| 130 |
|
| 131 |
-
|
| 132 |
|
| 133 |
#### Software
|
| 134 |
|
| 135 |
-
|
| 136 |
|
| 137 |
-
## Citation
|
| 138 |
|
| 139 |
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
| 140 |
|
| 141 |
**BibTeX:**
|
| 142 |
|
| 143 |
-
|
| 144 |
-
|
| 145 |
-
|
| 146 |
-
|
| 147 |
-
|
| 148 |
-
|
| 149 |
-
## Glossary [optional]
|
| 150 |
-
|
| 151 |
-
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
| 152 |
-
|
| 153 |
-
[More Information Needed]
|
| 154 |
-
|
| 155 |
-
## More Information [optional]
|
| 156 |
-
|
| 157 |
-
[More Information Needed]
|
| 158 |
-
|
| 159 |
-
## Model Card Authors [optional]
|
| 160 |
-
|
| 161 |
-
[More Information Needed]
|
| 162 |
-
|
| 163 |
-
## Model Card Contact
|
| 164 |
-
|
| 165 |
-
[More Information Needed]
|
|
|
|
| 36 |
|
| 37 |
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
| 38 |
|
| 39 |
+
The model can be run using LLaMA-Factory or any other framework capable of calling large language models. Users are recommended to follow the framework’s instructions for loading and running the model, ensuring sufficient GPU memory and compatibility with LoRA-adapted weights.
|
| 40 |
|
| 41 |
|
| 42 |
### Out-of-Scope Use
|
| 43 |
|
| 44 |
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
|
| 45 |
|
| 46 |
+
This model is specifically designed for perovskite solar cell precursor additives and is not intended for use in other fields. Using it outside this domain may produce unreliable or inaccurate suggestions, and could lead to incorrect conclusions if applied to unrelated materials or chemical systems.
|
| 47 |
|
| 48 |
## Bias, Risks, and Limitations
|
| 49 |
|
|
|
|
| 124 |
|
| 125 |
### Compute Infrastructure
|
| 126 |
|
| 127 |
+
Training was conducted on high-performance GPUs with bfloat16 precision, utilizing gradient accumulation and FlashAttention2 for efficiency. Users can run the model on GPUs with sufficient memory to accommodate LoRA-adapted weights.
|
| 128 |
|
| 129 |
#### Hardware
|
| 130 |
|
| 131 |
+
Training is performed on high-performance NVIDIA GPUs (e.g., H100 or equivalent). For inference, users should use GPUs with sufficient memory to load LoRA-adapted weights.
|
| 132 |
|
| 133 |
#### Software
|
| 134 |
|
| 135 |
+
The model was trained and can be run using Python (>=3.10), PyTorch, and LLaMA-Factory. Additional libraries such as Transformers and FlashAttention2 are used for optimized performance.
|
| 136 |
|
| 137 |
+
## Citation
|
| 138 |
|
| 139 |
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
| 140 |
|
| 141 |
**BibTeX:**
|
| 142 |
|
| 143 |
+
@article{wang2025perovskite,
|
| 144 |
+
title={Perovskite-R1: A Domain-Specialized LLM for Intelligent Discovery of Precursor Additives and Experimental Design},
|
| 145 |
+
author={Wang, Xin-De and Chen, Zhi-Rui and Guo, Peng-Jie and Gao, Ze-Feng and Mu, Cheng and Lu, Zhong-Yi},
|
| 146 |
+
journal={arXiv preprint arXiv:2507.16307},
|
| 147 |
+
year={2025}
|
| 148 |
+
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|