import gradio as gr import pandas as pd from sklearn.feature_extraction.text import TfidfVectorizer from sklearn.cluster import KMeans from sklearn.metrics import silhouette_score from sklearn.preprocessing import normalize import re from wordcloud import WordCloud import matplotlib.pyplot as plt def preprocess_data(df): df.rename(columns={'Question Asked': 'texts'}, inplace=True) df['texts'] = df['texts'].astype(str) df['texts'] = df['texts'].str.lower() df['texts'] = df['texts'].apply(lambda text: re.sub(r'https?://\S+|www\.\S+', '', text)) custom_synonyms = { 'application': ['form'], 'apply': ['fill', 'applied'], 'work': ['job'], 'salary': ['stipend', 'pay', 'payment', 'paid'], 'test': ['online test', 'amcat test', 'exam', 'assessment'], 'pass': ['clear', 'selected', 'pass or not'], 'result': ['outcome', 'mark', 'marks'], 'thanks': ["thanks a lot to you", "thankyou so much", "thank you so much", "tysm", "thank you", "okaythank", "thx", "ty", "thankyou", "thank", "thank u"], 'interview': ["pi"] } for original_word, synonym_list in custom_synonyms.items(): for synonym in synonym_list: pattern = r"\b" + synonym + r"\b(?!\s*\()" df['texts'] = df['texts'].str.replace(pattern, original_word, regex=True) pattern = r"\b" + synonym + r"\s+you" + r"\b(?!\s*\()" df['texts'] = df['texts'].str.replace(pattern, original_word + ' ', regex=True) spam_list = ["click here", "free", "recharge", "limited", "discount", "money back guarantee", "aaj", "kal", "mein", "how can i help you", "how can we help you", "how we can help you", "follow", "king", "contacting", "gar", "kirke", "subscribe", "youtube", "jio", "insta", "make money", "b2b","sent using truecaller"] rows_to_remove = set() for spam_phrase in spam_list: pattern = r"\b" + re.escape(spam_phrase) + r"\b" spam_rows = df['texts'].str.contains(pattern) rows_to_remove.update(df.index[spam_rows].tolist()) df = df.drop(rows_to_remove) greet_variations = ["hello", "hy", "hey", "hii", "hi", "heyyy", "bie", "bye"] for greet_var in greet_variations: pattern = r"(? 8 min)", "short reads (3-8 min)","not a student alumni","mock","share feedback","bite size (< 2 min)", "actually no","next steps","i'm a student alumni","i have questions"] for phrase in remove_phrases: df['texts'] = df['texts'].str.replace(phrase, '') general_variations = ["good morning", "good evening", "good afternoon", "good night", "done", "sorry", "top", "query", "stop", "sir", "sure", "oh", "wow", "aaa", "maam", "mam", "ma'am","i'm all set","ask a question","apply the survey", "videos (2-8 min)","long reads (> 8 min)","short reads (3-8 min)","not a student alumni","mock","share feedback","bite size (< 2 min)", "actually no","next steps","i'm a student alumni","i have questions"] for gen_var in general_variations: pattern = r"(?