Training in progress, epoch 1
Browse files- final/README.md +9 -13
- final/model.safetensors +1 -1
- model.safetensors +1 -1
- training_args.bin +1 -1
final/README.md
CHANGED
|
@@ -98,9 +98,9 @@ print(embeddings.shape)
|
|
| 98 |
# Get the similarity scores for the embeddings
|
| 99 |
similarities = model.similarity(embeddings, embeddings)
|
| 100 |
print(similarities)
|
| 101 |
-
# tensor([[1.0000, 0.
|
| 102 |
-
# [0.
|
| 103 |
-
# [0.
|
| 104 |
```
|
| 105 |
|
| 106 |
<!--
|
|
@@ -172,9 +172,9 @@ You can finetune this model on your own dataset.
|
|
| 172 |
|
| 173 |
- `per_device_train_batch_size`: 32
|
| 174 |
- `per_device_eval_batch_size`: 16
|
| 175 |
-
- `learning_rate`:
|
| 176 |
- `weight_decay`: 0.001
|
| 177 |
-
- `num_train_epochs`:
|
| 178 |
- `warmup_ratio`: 0.2
|
| 179 |
- `fp16`: True
|
| 180 |
- `dataloader_num_workers`: 2
|
|
@@ -199,13 +199,13 @@ You can finetune this model on your own dataset.
|
|
| 199 |
- `gradient_accumulation_steps`: 1
|
| 200 |
- `eval_accumulation_steps`: None
|
| 201 |
- `torch_empty_cache_steps`: None
|
| 202 |
-
- `learning_rate`:
|
| 203 |
- `weight_decay`: 0.001
|
| 204 |
- `adam_beta1`: 0.9
|
| 205 |
- `adam_beta2`: 0.999
|
| 206 |
- `adam_epsilon`: 1e-08
|
| 207 |
- `max_grad_norm`: 1.0
|
| 208 |
-
- `num_train_epochs`:
|
| 209 |
- `max_steps`: -1
|
| 210 |
- `lr_scheduler_type`: linear
|
| 211 |
- `lr_scheduler_kwargs`: {}
|
|
@@ -312,12 +312,8 @@ You can finetune this model on your own dataset.
|
|
| 312 |
| Epoch | Step | Training Loss |
|
| 313 |
|:------:|:----:|:-------------:|
|
| 314 |
| 0.0455 | 1 | 3.3947 |
|
| 315 |
-
| 1.0 | 22 | 2.
|
| 316 |
-
| 2.0 | 44 | 2.
|
| 317 |
-
| 3.0 | 66 | 2.6316 |
|
| 318 |
-
| 4.0 | 88 | 2.4885 |
|
| 319 |
-
| 5.0 | 110 | 2.4066 |
|
| 320 |
-
| 6.0 | 132 | 2.4311 |
|
| 321 |
|
| 322 |
|
| 323 |
### Framework Versions
|
|
|
|
| 98 |
# Get the similarity scores for the embeddings
|
| 99 |
similarities = model.similarity(embeddings, embeddings)
|
| 100 |
print(similarities)
|
| 101 |
+
# tensor([[1.0000, 0.8965, 0.4641],
|
| 102 |
+
# [0.8965, 1.0000, 0.4616],
|
| 103 |
+
# [0.4641, 0.4616, 1.0000]])
|
| 104 |
```
|
| 105 |
|
| 106 |
<!--
|
|
|
|
| 172 |
|
| 173 |
- `per_device_train_batch_size`: 32
|
| 174 |
- `per_device_eval_batch_size`: 16
|
| 175 |
+
- `learning_rate`: 2e-05
|
| 176 |
- `weight_decay`: 0.001
|
| 177 |
+
- `num_train_epochs`: 2
|
| 178 |
- `warmup_ratio`: 0.2
|
| 179 |
- `fp16`: True
|
| 180 |
- `dataloader_num_workers`: 2
|
|
|
|
| 199 |
- `gradient_accumulation_steps`: 1
|
| 200 |
- `eval_accumulation_steps`: None
|
| 201 |
- `torch_empty_cache_steps`: None
|
| 202 |
+
- `learning_rate`: 2e-05
|
| 203 |
- `weight_decay`: 0.001
|
| 204 |
- `adam_beta1`: 0.9
|
| 205 |
- `adam_beta2`: 0.999
|
| 206 |
- `adam_epsilon`: 1e-08
|
| 207 |
- `max_grad_norm`: 1.0
|
| 208 |
+
- `num_train_epochs`: 2
|
| 209 |
- `max_steps`: -1
|
| 210 |
- `lr_scheduler_type`: linear
|
| 211 |
- `lr_scheduler_kwargs`: {}
|
|
|
|
| 312 |
| Epoch | Step | Training Loss |
|
| 313 |
|:------:|:----:|:-------------:|
|
| 314 |
| 0.0455 | 1 | 3.3947 |
|
| 315 |
+
| 1.0 | 22 | 2.8004 |
|
| 316 |
+
| 2.0 | 44 | 2.4666 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 317 |
|
| 318 |
|
| 319 |
### Framework Versions
|
final/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 90864192
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7f4e1bfa7de51fde73187ce4b32fd4bcc39e11be28236be9ab0b752a3df6dd48
|
| 3 |
size 90864192
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 90864192
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:56aa0a763e86f0d86e91e4c4afcbf7fbae28a1d4c27cbbed96c0b10ec2118d07
|
| 3 |
size 90864192
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5752
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:849402676e03b666436dfa0358e1b7772cd390876fa060da707913822f48ffe5
|
| 3 |
size 5752
|