File size: 2,071 Bytes
e65dd55
d6a09f2
 
 
 
 
 
 
 
 
 
6711c73
 
 
46e8a7c
6711c73
 
 
d6a09f2
 
d85ef35
d6a09f2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6711c73
d6a09f2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e65dd55
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
import streamlit as st
import pandas as pd
import re
import tensorflow as tf
from nltk.corpus import stopwords
from nltk.tokenize import word_tokenize
from tensorflow.keras.preprocessing.text import Tokenizer
from tensorflow.keras.preprocessing.sequence import pad_sequences
from tensorflow.keras.models import load_model
import nltk

# Use /tmp for NLTK data (writable in Hugging Face Spaces)
nltk_data_dir = "/tmp/nltk_data"
nltk.data.path.append(nltk_data_dir)

# Download the stopwords and punkt resources
nltk.download('stopwords', download_dir=nltk_data_dir)
nltk.download('punkt', download_dir=nltk_data_dir)

# Load the trained model
model = load_model('src/model.keras')

# Load stopwords
stpwds_id = list(set(stopwords.words('english')))

# Text preprocessing function
def text_preprocessing(text):
    # Case folding
    text = text.lower()

    # Mention removal
    text = re.sub("@[A-Za-z0-9_]+", " ", text)

    # Hashtags removal
    text = re.sub("#[A-Za-z0-9_]+", " ", text)

    # Newline removal (\n)
    text = re.sub(r"\\n", " ",text)

    # Whitespace removal
    text = text.strip()

    # URL removal
    text = re.sub(r"http\S+", " ", text)
    text = re.sub(r"www.\S+", " ", text)

    # Non-letter removal (such as emoticons, symbols, etc.)
    text = re.sub("[^A-Za-z\s']", " ", text)

    # Tokenization
    tokens = word_tokenize(text)

    # Stopwords removal
    tokens = [word for word in tokens if word not in stpwds_id]

    # Combining Tokens
    text = ' '.join(tokens)

    return text

# Define the Streamlit interface
st.title('Sentiment Analysis App')

# Get user input
user_input = st.text_area("Enter the text for sentiment analysis:")

if st.button('Analyze'):
    if user_input:
        # Preprocess the input text
        processed_text = text_preprocessing(user_input)
        prediction = model.predict([[processed_text]])
        sentiment = "Positive" if prediction[0] > 0.5 else "Negative"

        # Display the result
        st.write(f"Sentiment: {sentiment}")
    else:
        st.write("Please enter some text.")