Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -5,7 +5,7 @@ def load_vectors(fname):
|
|
| 5 |
data = {}
|
| 6 |
for line in fin:
|
| 7 |
tokens = line.rstrip().split(' ')
|
| 8 |
-
data[tokens[0]] = tokens[1]
|
| 9 |
del fin
|
| 10 |
return data, sorted(data.keys(), key=len, reverse=True)
|
| 11 |
vectors, sorted_vector = load_vectors('wiki-news-300d-1M.vec')
|
|
@@ -41,8 +41,6 @@ class Trie:
|
|
| 41 |
return longest_match
|
| 42 |
|
| 43 |
def word2vec(word):
|
| 44 |
-
if not word in vectors:
|
| 45 |
-
return None
|
| 46 |
return list(vectors[word])
|
| 47 |
def tokenize(text):
|
| 48 |
trie = Trie()
|
|
|
|
| 5 |
data = {}
|
| 6 |
for line in fin:
|
| 7 |
tokens = line.rstrip().split(' ')
|
| 8 |
+
data[tokens[0]] = map(float, tokens[1:])
|
| 9 |
del fin
|
| 10 |
return data, sorted(data.keys(), key=len, reverse=True)
|
| 11 |
vectors, sorted_vector = load_vectors('wiki-news-300d-1M.vec')
|
|
|
|
| 41 |
return longest_match
|
| 42 |
|
| 43 |
def word2vec(word):
|
|
|
|
|
|
|
| 44 |
return list(vectors[word])
|
| 45 |
def tokenize(text):
|
| 46 |
trie = Trie()
|