saherPervaiz commited on
Commit
5775758
·
verified ·
1 Parent(s): 9aa6cc4

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +124 -208
app.py CHANGED
@@ -2,13 +2,13 @@ import streamlit as st
2
  import pandas as pd
3
  from sklearn.model_selection import train_test_split
4
  from sklearn.preprocessing import LabelEncoder
5
- from sklearn.ensemble import RandomForestClassifier
6
- from sklearn.linear_model import LogisticRegression
7
- from sklearn.svm import SVC
8
- from sklearn.neighbors import KNeighborsClassifier
9
- from sklearn.tree import DecisionTreeClassifier
10
  from sklearn.naive_bayes import GaussianNB
11
- from sklearn.metrics import accuracy_score, precision_score, recall_score, f1_score
12
  import numpy as np
13
  import matplotlib.pyplot as plt
14
  import seaborn as sns
@@ -24,7 +24,7 @@ if uploaded_file is not None:
24
  st.write("Dataset:")
25
  st.dataframe(df)
26
 
27
- # Convert categorical (str) data to numerical
28
  st.write("Converting Categorical Columns to Numerical Values:")
29
  label_encoder = LabelEncoder()
30
 
@@ -37,209 +37,125 @@ if uploaded_file is not None:
37
  st.write("Dataset After Conversion:")
38
  st.dataframe(df)
39
 
40
- # Model Training Section
41
- st.subheader("Model Training")
42
- if df.empty:
43
- st.warning("The dataset is empty. Please upload a valid CSV file.")
44
- else:
45
- # Handle Null Values (Missing Data)
46
- st.write("Handling Missing (Null) Values:")
47
- # Option to drop rows with null values or fill them
48
- fill_method = st.selectbox("Choose how to handle missing values", ["Drop rows", "Fill with mean/median"])
49
- if fill_method == "Drop rows":
50
- df = df.dropna()
51
- elif fill_method == "Fill with mean/median":
52
- for col in df.columns:
53
- if df[col].dtype in ['float64', 'int64']:
54
- df[col].fillna(df[col].mean(), inplace=True) # For numeric columns, fill with mean
55
- else:
56
- df[col].fillna(df[col].mode()[0], inplace=True) # For categorical columns, fill with mode
57
-
58
- # Handle Outliers using IQR method
59
- st.write("Handling Outliers:")
60
- # Define function to remove outliers using IQR
61
- def remove_outliers_iqr(dataframe):
62
- Q1 = dataframe.quantile(0.25)
63
- Q3 = dataframe.quantile(0.75)
64
- IQR = Q3 - Q1
65
- # Filter out rows that are outside the IQR range
66
- return dataframe[~((dataframe < (Q1 - 1.5 * IQR)) | (dataframe > (Q3 + 1.5 * IQR))).any(axis=1)]
67
-
68
- # Remove outliers from the numerical columns
69
- df = remove_outliers_iqr(df)
70
-
71
- # Handle Extreme Values by Capping (Winsorization)
72
- st.write("Handling Extreme Values (Capping):")
73
- def cap_extreme_values(dataframe):
74
- for col in dataframe.select_dtypes(include=[np.number]).columns:
75
- # Define the thresholds for extreme values (95th percentile and 5th percentile)
76
- lower_limit = dataframe[col].quantile(0.05)
77
- upper_limit = dataframe[col].quantile(0.95)
78
- # Cap the extreme values
79
- dataframe[col] = np.clip(dataframe[col], lower_limit, upper_limit)
80
- return dataframe
81
-
82
- df = cap_extreme_values(df)
83
-
84
- # Show cleaned dataset
85
- st.write("Cleaned Dataset:")
86
- st.dataframe(df)
87
 
88
- # Add clean data download option
89
- st.subheader("Download Cleaned Dataset")
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
90
 
91
- # Provide a download button for the cleaned dataset
92
  st.download_button(
93
- label="Download Cleaned Dataset (CSV)",
94
- data=df.to_csv(index=False), # Convert the cleaned dataset to CSV
95
- file_name="cleaned_dataset.csv", # Specify the file name
96
- mime="text/csv" # Specify the MIME type for CSV
97
  )
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
98
 
99
- target = st.selectbox("Select Target Variable", df.columns)
100
- features = [col for col in df.columns if col != target]
101
- X = df[features]
102
- y = df[target]
103
-
104
- # Label Encoding for categorical columns
105
- label_encoder = LabelEncoder()
106
-
107
- # Encode the target variable (if it's categorical)
108
- if y.dtype == 'object' or len(y.unique()) <= 10: # If the target variable is categorical
109
- y = label_encoder.fit_transform(y)
110
-
111
- # Encode categorical feature columns (if any)
112
- for col in X.columns:
113
- if X[col].dtype == 'object' or len(X[col].unique()) <= 10: # If the column is categorical
114
- X[col] = label_encoder.fit_transform(X[col])
115
-
116
- # Ensure there is enough data before proceeding with train-test split
117
- if len(X) == 0 or len(y) == 0:
118
- st.warning("Insufficient data. Please ensure there are valid feature and target columns.")
119
- else:
120
- # Split the data into training and test sets with customizable training size
121
- train_size = st.slider("Select Training Size", min_value=0.1, max_value=0.9, value=0.8)
122
- X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=1-train_size, random_state=42)
123
-
124
- # List of classifiers to evaluate
125
- classifiers = {
126
- 'Logistic Regression': LogisticRegression(max_iter=5000, solver='saga', penalty='l1'),
127
- 'Decision Tree': DecisionTreeClassifier(),
128
- 'Random Forest': RandomForestClassifier(),
129
- 'Support Vector Machine (SVM)': SVC(),
130
- 'K-Nearest Neighbors (k-NN)': KNeighborsClassifier(),
131
- 'Naive Bayes': GaussianNB()
132
- }
133
-
134
- # Initialize results storage
135
- metrics = []
136
-
137
- # Train and evaluate each model
138
- for name, classifier in classifiers.items():
139
- # Train the model
140
- classifier.fit(X_train, y_train)
141
-
142
- # Make predictions
143
- y_pred = classifier.predict(X_test)
144
-
145
- # Evaluate metrics
146
- accuracy = accuracy_score(y_test, y_pred)
147
- precision = precision_score(y_test, y_pred, zero_division=1, average='macro')
148
- recall = recall_score(y_test, y_pred, zero_division=1, average='macro')
149
- f1 = f1_score(y_test, y_pred, zero_division=1, average='macro')
150
-
151
- metrics.append({
152
- 'Model': name,
153
- 'Accuracy': round(accuracy, 2),
154
- 'Precision': round(precision, 2),
155
- 'Recall': round(recall, 2),
156
- 'F1-Score': round(f1, 2)
157
- })
158
-
159
- # Create a metrics DataFrame
160
- metrics_df = pd.DataFrame(metrics)
161
-
162
- # Display results in a table using st.dataframe
163
- st.subheader("Model Performance Metrics")
164
- st.dataframe(metrics_df)
165
-
166
- # Download options
167
- st.subheader("Download Model Performance Report in Different Formats")
168
-
169
- # CSV
170
- st.download_button(
171
- label="Download as CSV",
172
- data=metrics_df.to_csv(index=False),
173
- file_name="model_report.csv",
174
- mime="text/csv"
175
- )
176
-
177
- # JSON
178
- st.download_button(
179
- label="Download as JSON",
180
- data=metrics_df.to_json(orient='records'),
181
- file_name="model_report.json",
182
- mime="application/json"
183
- )
184
-
185
- # PDF (using `fpdf` library)
186
- from fpdf import FPDF
187
-
188
- def generate_pdf(df):
189
- pdf = FPDF()
190
- pdf.add_page()
191
- pdf.set_font("Arial", size=12)
192
- pdf.cell(200, 10, txt="Model Performance Report", ln=True, align="C")
193
- pdf.ln(10)
194
-
195
- # Add table header
196
- pdf.set_font("Arial", style='B', size=10)
197
- for header in df.columns:
198
- pdf.cell(40, 10, header, border=1)
199
- pdf.ln()
200
-
201
- # Add table rows
202
- pdf.set_font("Arial", size=10)
203
- for row in df.values:
204
- for value in row:
205
- pdf.cell(40, 10, str(value), border=1)
206
- pdf.ln()
207
-
208
- return pdf.output(dest='S').encode('latin1')
209
-
210
- # PDF download
211
- st.download_button(
212
- label="Download as PDF",
213
- data=generate_pdf(metrics_df),
214
- file_name="model_report.pdf",
215
- mime="application/pdf"
216
- )
217
-
218
-
219
- # Generate and download PNG report
220
- st.subheader("Download Report as PNG")
221
-
222
- # Create table plot using matplotlib
223
- fig, ax = plt.subplots(figsize=(12, 4)) # Adjust the figure size to match the table's layout
224
- ax.axis('tight')
225
- ax.axis('off')
226
- table_data = metrics_df.values
227
- table_columns = metrics_df.columns.tolist()
228
-
229
- table = ax.table(cellText=table_data, colLabels=table_columns, loc='center', cellLoc='center', colLoc='center')
230
- table.auto_set_font_size(False)
231
- table.set_fontsize(10)
232
- table.scale(1.2, 1.2) # Adjust the scale for better appearance
233
-
234
- # Save the table as a PNG file
235
- png_file = "model_report.png"
236
- fig.savefig(png_file, bbox_inches='tight', dpi=300)
237
-
238
- # Provide a download button for the PNG file
239
- with open(png_file, "rb") as file:
240
- st.download_button(
241
- label="Download as PNG",
242
- data=file,
243
- file_name="model_report.png",
244
- mime="image/png"
245
- )
 
2
  import pandas as pd
3
  from sklearn.model_selection import train_test_split
4
  from sklearn.preprocessing import LabelEncoder
5
+ from sklearn.ensemble import RandomForestClassifier, RandomForestRegressor
6
+ from sklearn.linear_model import LogisticRegression, LinearRegression
7
+ from sklearn.svm import SVC, SVR
8
+ from sklearn.neighbors import KNeighborsClassifier, KNeighborsRegressor
9
+ from sklearn.tree import DecisionTreeClassifier, DecisionTreeRegressor
10
  from sklearn.naive_bayes import GaussianNB
11
+ from sklearn.metrics import accuracy_score, precision_score, recall_score, f1_score, mean_squared_error, mean_absolute_error, r2_score
12
  import numpy as np
13
  import matplotlib.pyplot as plt
14
  import seaborn as sns
 
24
  st.write("Dataset:")
25
  st.dataframe(df)
26
 
27
+ # Convert categorical (str) data to numerical
28
  st.write("Converting Categorical Columns to Numerical Values:")
29
  label_encoder = LabelEncoder()
30
 
 
37
  st.write("Dataset After Conversion:")
38
  st.dataframe(df)
39
 
40
+ # Handle Null Values (Missing Data)
41
+ st.write("Handling Missing (Null) Values:")
42
+ fill_method = st.selectbox("Choose how to handle missing values", ["Drop rows", "Fill with mean/median"])
43
+ if fill_method == "Drop rows":
44
+ df = df.dropna()
45
+ elif fill_method == "Fill with mean/median":
46
+ for col in df.columns:
47
+ if df[col].dtype in ['float64', 'int64']:
48
+ df[col].fillna(df[col].mean(), inplace=True)
49
+ else:
50
+ df[col].fillna(df[col].mode()[0], inplace=True)
51
+
52
+ # Handle Outliers using IQR method
53
+ st.write("Handling Outliers:")
54
+ def remove_outliers_iqr(dataframe):
55
+ Q1 = dataframe.quantile(0.25)
56
+ Q3 = dataframe.quantile(0.75)
57
+ IQR = Q3 - Q1
58
+ return dataframe[~((dataframe < (Q1 - 1.5 * IQR)) | (dataframe > (Q3 + 1.5 * IQR))).any(axis=1)]
59
+
60
+ df = remove_outliers_iqr(df)
61
+
62
+ # Cap Extreme Values
63
+ st.write("Handling Extreme Values (Capping):")
64
+ def cap_extreme_values(dataframe):
65
+ for col in dataframe.select_dtypes(include=[np.number]).columns:
66
+ lower_limit = dataframe[col].quantile(0.05)
67
+ upper_limit = dataframe[col].quantile(0.95)
68
+ dataframe[col] = np.clip(dataframe[col], lower_limit, upper_limit)
69
+ return dataframe
70
+
71
+ df = cap_extreme_values(df)
72
+
73
+ # Show cleaned dataset
74
+ st.write("Cleaned Dataset:")
75
+ st.dataframe(df)
 
 
 
 
 
 
 
 
 
 
 
76
 
77
+ # Add clean data download option
78
+ st.subheader("Download Cleaned Dataset")
79
+ st.download_button(
80
+ label="Download Cleaned Dataset (CSV)",
81
+ data=df.to_csv(index=False),
82
+ file_name="cleaned_dataset.csv",
83
+ mime="text/csv"
84
+ )
85
+
86
+ target = st.selectbox("Select Target Variable", df.columns)
87
+ features = [col for col in df.columns if col != target]
88
+ X = df[features]
89
+ y = df[target]
90
+
91
+ if y.dtype == 'object' or len(y.unique()) <= 10: # Categorical target (classification)
92
+ st.subheader("Classification Model Training")
93
+ classifiers = {
94
+ 'Logistic Regression': LogisticRegression(max_iter=5000, solver='saga', penalty='l1'),
95
+ 'Decision Tree': DecisionTreeClassifier(),
96
+ 'Random Forest': RandomForestClassifier(),
97
+ 'Support Vector Machine (SVM)': SVC(),
98
+ 'K-Nearest Neighbors (k-NN)': KNeighborsClassifier(),
99
+ 'Naive Bayes': GaussianNB()
100
+ }
101
+
102
+ metrics = []
103
+ train_size = st.slider("Select Training Size", min_value=0.1, max_value=0.9, value=0.8)
104
+ X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=1-train_size, random_state=42)
105
+
106
+ for name, classifier in classifiers.items():
107
+ classifier.fit(X_train, y_train)
108
+ y_pred = classifier.predict(X_test)
109
+ metrics.append({
110
+ 'Model': name,
111
+ 'Accuracy': round(accuracy_score(y_test, y_pred), 2),
112
+ 'Precision': round(precision_score(y_test, y_pred, zero_division=1, average='macro'), 2),
113
+ 'Recall': round(recall_score(y_test, y_pred, zero_division=1, average='macro'), 2),
114
+ 'F1-Score': round(f1_score(y_test, y_pred, zero_division=1, average='macro'), 2)
115
+ })
116
+
117
+ metrics_df = pd.DataFrame(metrics)
118
+ st.subheader("Classification Model Performance Metrics")
119
+ st.dataframe(metrics_df)
120
 
 
121
  st.download_button(
122
+ label="Download Classification Report as CSV",
123
+ data=metrics_df.to_csv(index=False),
124
+ file_name="classification_report.csv",
125
+ mime="text/csv"
126
  )
127
+
128
+ else: # Continuous target (regression)
129
+ st.subheader("Regression Model Training")
130
+ regressors = {
131
+ 'Linear Regression': LinearRegression(),
132
+ 'Decision Tree Regressor': DecisionTreeRegressor(),
133
+ 'Random Forest Regressor': RandomForestRegressor(),
134
+ 'Support Vector Regressor (SVR)': SVR(),
135
+ 'K-Nearest Neighbors Regressor (k-NN)': KNeighborsRegressor()
136
+ }
137
+
138
+ regression_metrics = []
139
+ train_size = st.slider("Select Training Size", min_value=0.1, max_value=0.9, value=0.8)
140
+ X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=1-train_size, random_state=42)
141
+
142
+ for name, regressor in regressors.items():
143
+ regressor.fit(X_train, y_train)
144
+ y_pred = regressor.predict(X_test)
145
+ regression_metrics.append({
146
+ 'Model': name,
147
+ 'Mean Squared Error (MSE)': round(mean_squared_error(y_test, y_pred), 2),
148
+ 'Mean Absolute Error (MAE)': round(mean_absolute_error(y_test, y_pred), 2),
149
+ 'R² Score': round(r2_score(y_test, y_pred), 2)
150
+ })
151
+
152
+ regression_metrics_df = pd.DataFrame(regression_metrics)
153
+ st.subheader("Regression Model Performance Metrics")
154
+ st.dataframe(regression_metrics_df)
155
 
156
+ st.download_button(
157
+ label="Download Regression Report as CSV",
158
+ data=regression_metrics_df.to_csv(index=False),
159
+ file_name="regression_report.csv",
160
+ mime="text/csv"
161
+ )