hamzawaheed commited on
Commit
df3ea39
·
verified ·
1 Parent(s): e023819

Training in progress, epoch 1

Browse files
Files changed (3) hide show
  1. README.md +2 -27
  2. model.safetensors +1 -1
  3. training_args.bin +2 -2
README.md CHANGED
@@ -64,7 +64,7 @@ The model was trained and evaluated on the [`dair-ai/emotion`](https://huggingfa
64
 
65
  - **Total Samples:** 20,000
66
  - **Training Set:** 15,000 samples
67
- - **Validation Set:** 3,000 samples
68
  - **Test Set:** 2,000 samples
69
  - **Emotion Classes:** 6
70
  - **Joy:** 3,000 samples
@@ -104,18 +104,6 @@ The following hyperparameters were used during training:
104
  - **Gradient Accumulation Steps:** `2` (effectively simulating a batch size of `32`)
105
  - **Mixed Precision Training:** Enabled (Native AMP) if CUDA is available
106
 
107
- ### Training Environment
108
-
109
- - **Framework Versions:**
110
- - **Transformers:** `4.46.2`
111
- - **PyTorch:** `2.5.1+cu118`
112
- - **Datasets:** `3.1.0`
113
- - **Tokenizers:** `0.20.3`
114
- - **Hardware:**
115
- - **GPU:** NVIDIA RTX 3080 (Replace with your actual GPU)
116
- - **CPU:** 16-core Intel Xeon (Replace with your actual CPU)
117
- - **Memory:** 64 GB RAM (Replace with your actual RAM)
118
-
119
  ### Optimization Strategies
120
 
121
  - **Mixed Precision Training:** Utilized PyTorch's Native AMP to accelerate training and reduce memory consumption when a CUDA-enabled GPU is available.
@@ -125,10 +113,7 @@ The following hyperparameters were used during training:
125
 
126
  ### Training Duration
127
 
128
- - **Total Training Time:** Approximately `2.40` minutes (Note: Adjust this based on your actual training run)
129
- - **Training Steps:** `1000` steps
130
- - **Training Throughput:** `7.14` steps per second
131
-
132
  ### Logging and Monitoring
133
 
134
  - **Logging Directory:** `./logs`
@@ -142,13 +127,3 @@ After training, the model achieved the following performance metrics:
142
 
143
  - **Validation Accuracy:** `93.10%`
144
  - **Test Accuracy:** `93.10%`
145
-
146
- *Note:* The identical validation and test accuracies suggest consistent performance across both datasets. It's recommended to further investigate to ensure data splits are correctly managed and to explore additional metrics for a more comprehensive evaluation.
147
-
148
- ### Framework Versions
149
-
150
- - **Transformers:** `4.46.2`
151
- - **PyTorch:** `2.5.1+cu118`
152
- - **Datasets:** `3.1.0`
153
- - **Tokenizers:** `0.20.3`
154
-
 
64
 
65
  - **Total Samples:** 20,000
66
  - **Training Set:** 15,000 samples
67
+ - **Validation Set:** 2,000 samples
68
  - **Test Set:** 2,000 samples
69
  - **Emotion Classes:** 6
70
  - **Joy:** 3,000 samples
 
104
  - **Gradient Accumulation Steps:** `2` (effectively simulating a batch size of `32`)
105
  - **Mixed Precision Training:** Enabled (Native AMP) if CUDA is available
106
 
 
 
 
 
 
 
 
 
 
 
 
 
107
  ### Optimization Strategies
108
 
109
  - **Mixed Precision Training:** Utilized PyTorch's Native AMP to accelerate training and reduce memory consumption when a CUDA-enabled GPU is available.
 
113
 
114
  ### Training Duration
115
 
116
+ - **Total Training Time:** Approximately `2.40` minutes
 
 
 
117
  ### Logging and Monitoring
118
 
119
  - **Logging Directory:** `./logs`
 
127
 
128
  - **Validation Accuracy:** `93.10%`
129
  - **Test Accuracy:** `93.10%`
 
 
 
 
 
 
 
 
 
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:58ad59989e79613007578c6a60e0f4b8e34b0a8b345b346b83221148ece02476
3
  size 267844872
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19d1a3c44772e96b5327f31056e7ac16e6cefd360b89807853755221ba43d3a5
3
  size 267844872
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e00c744292e2d5af6d7b83fd84402d005f758ef841400848da9b49d9e7b1bded
3
- size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1f16fa5596c73967f5afc8c25e593f01bfb06ae91ebe31f9157c543ed6f4ef1
3
+ size 5240