blainetrain commited on
Commit
5d020bd
·
verified ·
1 Parent(s): 0335f64

End of training

Browse files
Files changed (2) hide show
  1. README.md +50 -54
  2. model.safetensors +1 -1
README.md CHANGED
@@ -1,72 +1,68 @@
1
  ---
2
- license: apache-2.0
3
- tags:
4
- - chemistry
5
- - precite
6
- - chemberta
7
- datasets:
8
- - blainetrain/precite-dataset-FLP-Test-v11
9
  base_model: seyonec/ChemBERTa-zinc-base-v1
 
 
 
 
 
 
 
10
  model-index:
11
- - name: FLP-Test-v11
12
- results:
13
- - task:
14
- type: molecular-property-prediction
15
- metrics:
16
- - name: Accuracy
17
- type: accuracy
18
- value: 0.5000
19
- - name: F1
20
- type: f1
21
- value: 0.5000
22
- - name: Precision
23
- type: precision
24
- value: 0.5000
25
- - name: Recall
26
- type: recall
27
- value: 0.5000
28
  ---
29
 
30
- # FLP Test v11
 
 
 
 
 
 
 
 
 
 
 
 
 
31
 
32
- A chemistry prediction model fine-tuned on Precite platform.
33
 
34
- ## Model Details
35
 
36
- - **Base Model**: [seyonec/ChemBERTa-zinc-base-v1](https://huggingface.co/seyonec/ChemBERTa-zinc-base-v1)
37
- - **Fine-tuned On**: 8 training samples, 2 validation samples (80/20 split)
38
- - **Task**: Molecular property prediction (4 classes)
39
- - **Epochs**: 2
40
- - **Training Date**: 2026-02-04
41
 
42
- ## Performance Metrics (20% Holdout Test Set)
43
 
44
- | Metric | Value |
45
- |--------|-------|
46
- | **Accuracy** | 0.5000 |
47
- | **F1 Score** | 0.5000 |
48
- | **Precision** | 0.5000 |
49
- | **Recall** | 0.5000 |
50
- | Training Loss | 1.3901 |
51
 
52
- ## Label Classes
53
 
54
- - `high`
55
- - `low`
56
- - `medium`
57
- - `very_low`
58
 
59
- ## Usage
 
 
 
 
 
 
 
 
60
 
61
- This model can be queried through the Precite platform for FLP chemistry predictions.
62
 
63
- ```python
64
- from transformers import AutoModelForSequenceClassification, AutoTokenizer
 
 
65
 
66
- model = AutoModelForSequenceClassification.from_pretrained("blainetrain/FLP-Test-v11")
67
- tokenizer = AutoTokenizer.from_pretrained("blainetrain/FLP-Test-v11")
68
- ```
69
 
70
- ## Training Data
71
 
72
- See the associated dataset: [blainetrain/precite-dataset-FLP-Test-v11](https://huggingface.co/datasets/blainetrain/precite-dataset-FLP-Test-v11)
 
 
 
 
1
  ---
2
+ library_name: transformers
 
 
 
 
 
 
3
  base_model: seyonec/ChemBERTa-zinc-base-v1
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
+ - precision
9
+ - recall
10
+ - f1
11
  model-index:
12
+ - name: FLP-Test-v11
13
+ results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  ---
15
 
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+ # FLP-Test-v11
20
+
21
+ This model is a fine-tuned version of [seyonec/ChemBERTa-zinc-base-v1](https://huggingface.co/seyonec/ChemBERTa-zinc-base-v1) on the None dataset.
22
+ It achieves the following results on the evaluation set:
23
+ - Loss: 1.3951
24
+ - Accuracy: 0.0
25
+ - Precision: 0.0
26
+ - Recall: 0.0
27
+ - F1: 0.0
28
+
29
+ ## Model description
30
 
31
+ More information needed
32
 
33
+ ## Intended uses & limitations
34
 
35
+ More information needed
 
 
 
 
36
 
37
+ ## Training and evaluation data
38
 
39
+ More information needed
 
 
 
 
 
 
40
 
41
+ ## Training procedure
42
 
43
+ ### Training hyperparameters
 
 
 
44
 
45
+ The following hyperparameters were used during training:
46
+ - learning_rate: 5e-05
47
+ - train_batch_size: 4
48
+ - eval_batch_size: 4
49
+ - seed: 42
50
+ - optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
51
+ - lr_scheduler_type: linear
52
+ - lr_scheduler_warmup_steps: 100
53
+ - num_epochs: 2
54
 
55
+ ### Training results
56
 
57
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 |
58
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:---:|
59
+ | No log | 1.0 | 2 | 1.3951 | 0.0 | 0.0 | 0.0 | 0.0 |
60
+ | No log | 2.0 | 4 | 1.3983 | 0.0 | 0.0 | 0.0 | 0.0 |
61
 
 
 
 
62
 
63
+ ### Framework versions
64
 
65
+ - Transformers 5.0.0
66
+ - Pytorch 2.10.0+cu128
67
+ - Datasets 4.5.0
68
+ - Tokenizers 0.22.2
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:78066d95624cdde21e80bca5026c624681a64b793d4d1859fdca7d86177a31cd
3
  size 176440432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ca2d2d62cb6afe5944335ac67892f0cc2b83cde2a6abd50a6fb70461ed06e4a
3
  size 176440432