JH976 commited on
Commit
f995437
·
1 Parent(s): 307a8cc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -10
README.md CHANGED
@@ -68,26 +68,20 @@ Use the model as a reference for research support and hypothesis generation in p
68
 
69
  <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
70
 
71
- [More Information Needed]
72
 
73
  ### Training Procedure
74
 
75
  <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
76
 
77
- #### Preprocessing [optional]
78
-
79
- [More Information Needed]
80
-
81
 
82
  #### Training Hyperparameters
83
 
84
  - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
85
 
86
- #### Speeds, Sizes, Times [optional]
87
-
88
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
89
-
90
- [More Information Needed]
91
 
92
  ## Evaluation
93
 
 
68
 
69
  <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
70
 
71
+ The model was trained on a curated collection of scientific literature, experimental datasets, and publicly available resources related to perovskite solar cell precursor additives. The dataset includes research articles and drug databases, focusing on synthesis, additive effects, and device performance. All training data has been uploaded and is documented for transparency and reproducibility.
72
 
73
  ### Training Procedure
74
 
75
  <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
76
 
77
+ The model was trained using a transformer-based architecture optimized for scientific text.
78
+ Training was performed on high-performance GPUs with gradient accumulation.
79
+ Fine-tuning was conducted on curated perovskite precursor additive datasets.
 
80
 
81
  #### Training Hyperparameters
82
 
83
  - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
84
 
 
 
 
 
 
85
 
86
  ## Evaluation
87