Spaces:
Running
Running
changes
Browse files- .gitattributes +2 -0
- model.py +5 -5
- saved_models/vocab.file +0 -0
- saved_vocabulary/vocab_1.file +3 -0
- saved_vocabulary/vocab_2.file +3 -0
- tmp.jpg +0 -0
.gitattributes
CHANGED
|
@@ -29,3 +29,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 29 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 30 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 31 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
| 29 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 30 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 31 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 32 |
+
*.file filter=lfs diff=lfs merge=lfs -text
|
| 33 |
+
vocab_2.file filter=lfs diff=lfs merge=lfs -text
|
model.py
CHANGED
|
@@ -6,7 +6,7 @@ import numpy as np
|
|
| 6 |
|
| 7 |
# CONTANTS
|
| 8 |
MAX_LENGTH = 40
|
| 9 |
-
VOCABULARY_SIZE = 10000
|
| 10 |
BATCH_SIZE = 32
|
| 11 |
BUFFER_SIZE = 1000
|
| 12 |
EMBEDDING_DIM = 512
|
|
@@ -14,10 +14,10 @@ UNITS = 512
|
|
| 14 |
|
| 15 |
|
| 16 |
# LOADING DATA
|
| 17 |
-
vocab = pickle.load(open('
|
| 18 |
|
| 19 |
tokenizer = tf.keras.layers.TextVectorization(
|
| 20 |
-
max_tokens=VOCABULARY_SIZE,
|
| 21 |
standardize=None,
|
| 22 |
output_sequence_length=MAX_LENGTH,
|
| 23 |
vocabulary=vocab
|
|
@@ -321,8 +321,8 @@ def get_caption_model():
|
|
| 321 |
caption_model.decoder(sample_y, sample_enc_out, training=False)
|
| 322 |
|
| 323 |
try:
|
| 324 |
-
caption_model.load_weights('saved_models/
|
| 325 |
except FileNotFoundError:
|
| 326 |
-
caption_model.load_weights('Image-Captioning/saved_models/
|
| 327 |
|
| 328 |
return caption_model
|
|
|
|
| 6 |
|
| 7 |
# CONTANTS
|
| 8 |
MAX_LENGTH = 40
|
| 9 |
+
# VOCABULARY_SIZE = 10000
|
| 10 |
BATCH_SIZE = 32
|
| 11 |
BUFFER_SIZE = 1000
|
| 12 |
EMBEDDING_DIM = 512
|
|
|
|
| 14 |
|
| 15 |
|
| 16 |
# LOADING DATA
|
| 17 |
+
vocab = pickle.load(open('saved_vocabulary/vocab_1.file', 'rb'))
|
| 18 |
|
| 19 |
tokenizer = tf.keras.layers.TextVectorization(
|
| 20 |
+
# max_tokens=VOCABULARY_SIZE,
|
| 21 |
standardize=None,
|
| 22 |
output_sequence_length=MAX_LENGTH,
|
| 23 |
vocabulary=vocab
|
|
|
|
| 321 |
caption_model.decoder(sample_y, sample_enc_out, training=False)
|
| 322 |
|
| 323 |
try:
|
| 324 |
+
caption_model.load_weights('saved_models/image_captioning_transformer_weights_2.h5')
|
| 325 |
except FileNotFoundError:
|
| 326 |
+
caption_model.load_weights('Image-Captioning/saved_models/image_captioning_transformer_weights_2.h5')
|
| 327 |
|
| 328 |
return caption_model
|
saved_models/vocab.file
DELETED
|
Binary file (860 kB)
|
|
|
saved_vocabulary/vocab_1.file
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4df22d5dd84bbb3421dfce39b4affd49ba2dc08739627ee9cf036ae17d326d28
|
| 3 |
+
size 860291
|
saved_vocabulary/vocab_2.file
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5397f3824177102eff3f2fe01eb8b6fffbe97671a8e8b102e20bd5e8e2064b7e
|
| 3 |
+
size 1985473
|
tmp.jpg
CHANGED
|
|