Training in progress, epoch 0
Browse files- config.json +8 -4
- pytorch_model.bin +1 -1
config.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
{
|
| 2 |
-
"_commit_hash":
|
| 3 |
-
"_name_or_path": "
|
| 4 |
"architectures": [
|
| 5 |
"VisionEncoderDecoderModel"
|
| 6 |
],
|
|
@@ -91,6 +91,10 @@
|
|
| 91 |
},
|
| 92 |
"decoder_start_token_id": 57547,
|
| 93 |
"encoder": {
|
|
|
|
|
|
|
|
|
|
|
|
|
| 94 |
"_name_or_path": "",
|
| 95 |
"add_cross_attention": false,
|
| 96 |
"architectures": null,
|
|
@@ -126,8 +130,8 @@
|
|
| 126 |
"1": "LABEL_1"
|
| 127 |
},
|
| 128 |
"image_size": [
|
| 129 |
-
|
| 130 |
-
|
| 131 |
],
|
| 132 |
"initializer_range": 0.02,
|
| 133 |
"is_decoder": false,
|
|
|
|
| 1 |
{
|
| 2 |
+
"_commit_hash": null,
|
| 3 |
+
"_name_or_path": "/models/model",
|
| 4 |
"architectures": [
|
| 5 |
"VisionEncoderDecoderModel"
|
| 6 |
],
|
|
|
|
| 91 |
},
|
| 92 |
"decoder_start_token_id": 57547,
|
| 93 |
"encoder": {
|
| 94 |
+
"IMAGE_SIZE": [
|
| 95 |
+
1280,
|
| 96 |
+
960
|
| 97 |
+
],
|
| 98 |
"_name_or_path": "",
|
| 99 |
"add_cross_attention": false,
|
| 100 |
"architectures": null,
|
|
|
|
| 130 |
"1": "LABEL_1"
|
| 131 |
},
|
| 132 |
"image_size": [
|
| 133 |
+
2560,
|
| 134 |
+
1920
|
| 135 |
],
|
| 136 |
"initializer_range": 0.02,
|
| 137 |
"is_decoder": false,
|
pytorch_model.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 809273113
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e5c4528d3dbc605328063937ec655493fa88fd39f87b1ec8b90d03fca9e2dfaf
|
| 3 |
size 809273113
|