ebhon commited on
Commit
1ac510e
·
verified ·
1 Parent(s): c8cfc38

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -22
app.py CHANGED
@@ -13,28 +13,8 @@ from nltk.stem import WordNetLemmatizer
13
  from nltk.tokenize import word_tokenize, sent_tokenize
14
  from sklearn.feature_extraction.text import CountVectorizer
15
 
16
- nltk_data_path = '/home/user/nltk_data'
17
-
18
-
19
- if not os.path.exists(nltk_data_path):
20
- os.makedirs(nltk_data_path)
21
-
22
-
23
- nltk.data.path.append(nltk_data_path)
24
- nltk.download('punkt', download_dir=nltk_data_path)
25
- nltk.download('stopwords', download_dir=nltk_data_path)
26
-
27
-
28
- from nltk.corpus import stopwords
29
- from nltk.tokenize import sent_tokenize
30
-
31
-
32
- stop_words = set(stopwords.words('english'))
33
-
34
-
35
- sample_text = "This is a test sentence. Let's see if it works!"
36
- sentences = sent_tokenize(sample_text)
37
- print(sentences)
38
 
39
  model = load_model('model_improved.keras')
40
  vectorizer = joblib.load('vectorizer.joblib')
@@ -44,6 +24,7 @@ with open('product_mapping.json', 'r') as file1:
44
  reverse_mapping = {v: k for k, v in product_mapping.items()}
45
 
46
  lemmatizer = WordNetLemmatizer()
 
47
 
48
  def clean_text(text):
49
  if text is None:
 
13
  from nltk.tokenize import word_tokenize, sent_tokenize
14
  from sklearn.feature_extraction.text import CountVectorizer
15
 
16
+ nltk.download('stopwords')
17
+ nltk.download('punkt')
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
 
19
  model = load_model('model_improved.keras')
20
  vectorizer = joblib.load('vectorizer.joblib')
 
24
  reverse_mapping = {v: k for k, v in product_mapping.items()}
25
 
26
  lemmatizer = WordNetLemmatizer()
27
+ stop_words = set(stopwords.words('english'))
28
 
29
  def clean_text(text):
30
  if text is None: