Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -28,8 +28,16 @@ with open('speechesdataset/pres_dict.pkl', 'rb') as file1:
|
|
| 28 |
reversed_dict = pickle.load(file1)
|
| 29 |
pres_dict = {value: key for key, value in reversed_dict.items()}
|
| 30 |
|
| 31 |
-
with open('speechesdataset/
|
| 32 |
-
tokenizer = pickle.load(file)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
|
| 34 |
# load in model
|
| 35 |
model = Classifier(tokenizer.vocab_size)
|
|
|
|
| 28 |
reversed_dict = pickle.load(file1)
|
| 29 |
pres_dict = {value: key for key, value in reversed_dict.items()}
|
| 30 |
|
| 31 |
+
with open('speechesdataset/tokenizer_stoi.pkl', 'rb') as file:
|
| 32 |
+
tokenizer.stoi = pickle.load(file)
|
| 33 |
+
|
| 34 |
+
with open('speechesdataset/tokenizer_itos.pkl', 'rb') as file:
|
| 35 |
+
tokenizer.itos = pickle.load(file)
|
| 36 |
+
|
| 37 |
+
with open('speechesdataset/tokenizer_vocab.pkl', 'rb') as file:
|
| 38 |
+
tokenizer.vocab = pickle.load(file)
|
| 39 |
+
|
| 40 |
+
tokenizer.vocab_size = len(tokenizer.stoi)
|
| 41 |
|
| 42 |
# load in model
|
| 43 |
model = Classifier(tokenizer.vocab_size)
|