Update README.md
Browse files
README.md
CHANGED
|
@@ -68,26 +68,20 @@ Use the model as a reference for research support and hypothesis generation in p
|
|
| 68 |
|
| 69 |
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
| 70 |
|
| 71 |
-
|
| 72 |
|
| 73 |
### Training Procedure
|
| 74 |
|
| 75 |
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
| 76 |
|
| 77 |
-
|
| 78 |
-
|
| 79 |
-
|
| 80 |
-
|
| 81 |
|
| 82 |
#### Training Hyperparameters
|
| 83 |
|
| 84 |
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
| 85 |
|
| 86 |
-
#### Speeds, Sizes, Times [optional]
|
| 87 |
-
|
| 88 |
-
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
| 89 |
-
|
| 90 |
-
[More Information Needed]
|
| 91 |
|
| 92 |
## Evaluation
|
| 93 |
|
|
|
|
| 68 |
|
| 69 |
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
| 70 |
|
| 71 |
+
The model was trained on a curated collection of scientific literature, experimental datasets, and publicly available resources related to perovskite solar cell precursor additives. The dataset includes research articles and drug databases, focusing on synthesis, additive effects, and device performance. All training data has been uploaded and is documented for transparency and reproducibility.
|
| 72 |
|
| 73 |
### Training Procedure
|
| 74 |
|
| 75 |
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
| 76 |
|
| 77 |
+
The model was trained using a transformer-based architecture optimized for scientific text.
|
| 78 |
+
Training was performed on high-performance GPUs with gradient accumulation.
|
| 79 |
+
Fine-tuning was conducted on curated perovskite precursor additive datasets.
|
|
|
|
| 80 |
|
| 81 |
#### Training Hyperparameters
|
| 82 |
|
| 83 |
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
| 84 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 85 |
|
| 86 |
## Evaluation
|
| 87 |
|