LamaDiab commited on
Commit
f874e41
·
verified ·
1 Parent(s): 2d66003

Training in progress, epoch 1

Browse files
final/README.md CHANGED
@@ -98,9 +98,9 @@ print(embeddings.shape)
98
  # Get the similarity scores for the embeddings
99
  similarities = model.similarity(embeddings, embeddings)
100
  print(similarities)
101
- # tensor([[1.0000, 0.8937, 0.4770],
102
- # [0.8937, 1.0000, 0.4732],
103
- # [0.4770, 0.4732, 1.0000]])
104
  ```
105
 
106
  <!--
@@ -172,9 +172,9 @@ You can finetune this model on your own dataset.
172
 
173
  - `per_device_train_batch_size`: 32
174
  - `per_device_eval_batch_size`: 16
175
- - `learning_rate`: 1e-05
176
  - `weight_decay`: 0.001
177
- - `num_train_epochs`: 6
178
  - `warmup_ratio`: 0.2
179
  - `fp16`: True
180
  - `dataloader_num_workers`: 2
@@ -199,13 +199,13 @@ You can finetune this model on your own dataset.
199
  - `gradient_accumulation_steps`: 1
200
  - `eval_accumulation_steps`: None
201
  - `torch_empty_cache_steps`: None
202
- - `learning_rate`: 1e-05
203
  - `weight_decay`: 0.001
204
  - `adam_beta1`: 0.9
205
  - `adam_beta2`: 0.999
206
  - `adam_epsilon`: 1e-08
207
  - `max_grad_norm`: 1.0
208
- - `num_train_epochs`: 6
209
  - `max_steps`: -1
210
  - `lr_scheduler_type`: linear
211
  - `lr_scheduler_kwargs`: {}
@@ -312,12 +312,8 @@ You can finetune this model on your own dataset.
312
  | Epoch | Step | Training Loss |
313
  |:------:|:----:|:-------------:|
314
  | 0.0455 | 1 | 3.3947 |
315
- | 1.0 | 22 | 2.9755 |
316
- | 2.0 | 44 | 2.8228 |
317
- | 3.0 | 66 | 2.6316 |
318
- | 4.0 | 88 | 2.4885 |
319
- | 5.0 | 110 | 2.4066 |
320
- | 6.0 | 132 | 2.4311 |
321
 
322
 
323
  ### Framework Versions
 
98
  # Get the similarity scores for the embeddings
99
  similarities = model.similarity(embeddings, embeddings)
100
  print(similarities)
101
+ # tensor([[1.0000, 0.8965, 0.4641],
102
+ # [0.8965, 1.0000, 0.4616],
103
+ # [0.4641, 0.4616, 1.0000]])
104
  ```
105
 
106
  <!--
 
172
 
173
  - `per_device_train_batch_size`: 32
174
  - `per_device_eval_batch_size`: 16
175
+ - `learning_rate`: 2e-05
176
  - `weight_decay`: 0.001
177
+ - `num_train_epochs`: 2
178
  - `warmup_ratio`: 0.2
179
  - `fp16`: True
180
  - `dataloader_num_workers`: 2
 
199
  - `gradient_accumulation_steps`: 1
200
  - `eval_accumulation_steps`: None
201
  - `torch_empty_cache_steps`: None
202
+ - `learning_rate`: 2e-05
203
  - `weight_decay`: 0.001
204
  - `adam_beta1`: 0.9
205
  - `adam_beta2`: 0.999
206
  - `adam_epsilon`: 1e-08
207
  - `max_grad_norm`: 1.0
208
+ - `num_train_epochs`: 2
209
  - `max_steps`: -1
210
  - `lr_scheduler_type`: linear
211
  - `lr_scheduler_kwargs`: {}
 
312
  | Epoch | Step | Training Loss |
313
  |:------:|:----:|:-------------:|
314
  | 0.0455 | 1 | 3.3947 |
315
+ | 1.0 | 22 | 2.8004 |
316
+ | 2.0 | 44 | 2.4666 |
 
 
 
 
317
 
318
 
319
  ### Framework Versions
final/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f35e9fc17b1dead17057de429ab6bae16262ea5718de8465be0dc823cf9b7c78
3
  size 90864192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f4e1bfa7de51fde73187ce4b32fd4bcc39e11be28236be9ab0b752a3df6dd48
3
  size 90864192
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f4e1bfa7de51fde73187ce4b32fd4bcc39e11be28236be9ab0b752a3df6dd48
3
  size 90864192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56aa0a763e86f0d86e91e4c4afcbf7fbae28a1d4c27cbbed96c0b10ec2118d07
3
  size 90864192
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:92d5a20586e1e0b0b585e149a74b68da5abb96f10841c23444985b42d5be4961
3
  size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:849402676e03b666436dfa0358e1b7772cd390876fa060da707913822f48ffe5
3
  size 5752