| import pandas as pd |
| import numpy as np |
| import matplotlib.pyplot as plt |
| import time |
| import seaborn as sns |
| import warnings |
|
|
| warnings.filterwarnings('ignore') |
|
|
| data = pd.read_csv('/content/Credit_Data.csv') |
|
|
| data.head() |
|
|
| data.drop('ID',axis=1,inplace=True) |
|
|
| data.shape |
|
|
| def get_summary(df): |
| df_desc = pd.DataFrame(df.describe(include='all').transpose()) |
| df_summary = pd.DataFrame({ |
| 'dtype': df.dtypes, |
| '#missing': df.isnull().sum().values, |
| '#duplicates': df.duplicated().sum(), |
| '#unique': df.nunique().values, |
| 'min': df_desc['min'].values, |
| 'max': df_desc['max'].values, |
| 'avg': df_desc['mean'].values, |
| 'std dev': df_desc['std'].values, |
| }) |
| return df_summary |
|
|
| get_summary(data).style.background_gradient() |
|
|
| target_col = 'Balance' |
| feature = data.drop('Balance', axis=1).columns |
|
|
| fig, ax = plt.subplots(2, 5, figsize=(20, 10)) |
| axes = ax.flatten() |
|
|
| for i, col in enumerate(data[feature].columns): |
| sns.scatterplot(data=data, x=col, y='Balance', hue='Gender', ax=axes[i]) |
|
|
| fig.suptitle('Interactions between Target Column and Features') |
| plt.tight_layout() |
| plt.show() |
|
|
| fig, ax = plt.subplots(2, 6, figsize=(20, 10)) |
| axes = ax.flatten() |
|
|
| for i, col in enumerate(data.columns): |
| sns.histplot(data=data, x=col, hue='Gender', ax=axes[i]) |
|
|
| fig.suptitle("Gender-Based Distribution of Financial and Demographic Features in the Dataset") |
| plt.tight_layout() |
|
|
| for ax in axes: |
| if not ax.has_data(): |
| fig.delaxes(ax) |
|
|
| plt.show() |
|
|
| sns.pairplot(data, kind='scatter', diag_kind='hist', hue='Gender', palette='colorblind') |
|
|
| numeric_columns = data.select_dtypes(include='number').columns |
|
|
| fig, ax = plt.subplots(len(numeric_columns), 2, figsize=(12, len(numeric_columns)*2)) |
| ax = ax.flatten() |
|
|
| for i, col in enumerate(numeric_columns): |
| sns.boxplot(data=data, x=col, width=0.6, ax=ax[2*i]) |
| sns.violinplot(data=data, x=col, ax=ax[2*i + 1]) |
|
|
| plt.tight_layout() |
| plt.show() |
|
|
| corr = data.select_dtypes(exclude='object').corr(method='spearman') |
| mask = np.triu(np.ones_like(corr)) |
|
|
| sns.heatmap(corr, annot=True, mask=mask, cmap='YlGnBu',cbar=True) |
| plt.title('Correlation Matrix',fontdict={'color': 'blue', 'fontsize': 12}) |
|
|
| from sklearn.preprocessing import OneHotEncoder |
|
|
| cat_columns = data.select_dtypes(include='O').columns.to_list() |
|
|
| dummie_df = pd.get_dummies(data=data[cat_columns], drop_first=True).astype('int8') |
|
|
| df = data.join(dummie_df) |
| df.drop(cat_columns,axis=1,inplace=True) |
|
|
| df.head() |
|
|
| from imblearn.over_sampling import SMOTE |
| from collections import Counter |
|
|
| X_train = df.drop('Student_Yes',axis=1) |
| y_train = df['Student_Yes'] |
|
|
| sm = SMOTE(sampling_strategy='minority',random_state=14, k_neighbors=5, n_jobs=-1) |
| sm_X_train, sm_Y_train = sm.fit_resample(X_train,y_train) |
|
|
| print('Before sampling class distribution', Counter(y_train)) |
| print('\nAfter sampling class distribution', Counter(sm_Y_train)) |
|
|
| sm_df = pd.concat([sm_X_train,sm_Y_train],axis=1) |
| sm_df.head() |
|
|
| get_summary(sm_df).style.background_gradient() |
|
|
| !pip install ydata_profiling |
|
|
| from ydata_profiling import ProfileReport |
|
|
| profile_report = ProfileReport( |
| sm_df, |
| sort=None, |
| progress_bar=False, |
| html = {'style': {'full_width': True}}, |
| correlations={ |
| "auto": {"calculate": True}, |
| "pearson": {"calculate": False}, |
| "spearman": {"calculate": False}, |
| "kendall": {"calculate": False}, |
| "phi_k": {"calculate": True}, |
| "cramers": {"calculate": True}, |
| }, |
| explorative=True, |
| title="Profiling Report" |
| ) |
|
|
| profile_report.to_file('output.html') |
|
|
| from sklearn.linear_model import LinearRegression |
| from sklearn.model_selection import train_test_split, cross_val_score |
| from sklearn.preprocessing import StandardScaler |
| from sklearn import metrics |
|
|
| X = sm_df.drop('Balance',axis=1) |
| y = sm_df.Balance |
|
|
| train_x, valid_x, train_y, valid_y = train_test_split(X,y, test_size=0.2, random_state=16518, shuffle=True) |
|
|
| scaler = StandardScaler() |
| train_x = scaler.fit_transform(train_x) |
| valid_x = scaler.transform(valid_x) |
|
|
| lm = LinearRegression() |
| history = lm.fit(train_x, train_y) |
| pred = lm.predict(valid_x) |
| r2 = metrics.r2_score(valid_y,pred) |
|
|
| print('r2_score',r2) |
|
|
| lm_df = pd.DataFrame(history.coef_.T, index= X.columns, columns=['coef_']) |
|
|
| lm_df.loc['intercept_'] = lm.intercept_ |
|
|
| lm_df.sort_values(by='coef_') |
|
|
| plt.barh(y= lm_df.index, width='coef_', data=lm_df) |
| plt.show() |
|
|
| from sklearn.model_selection import train_test_split, cross_val_score, KFold |
| from sklearn.preprocessing import StandardScaler |
| from sklearn.preprocessing import PolynomialFeatures |
| from sklearn import metrics |
|
|
| X = sm_df.drop('Balance',axis=1) |
| y = sm_df.Balance |
|
|
| train_x, valid_x, train_y, valid_y = train_test_split(X,y, test_size=0.2, random_state=16518, shuffle=True) |
|
|
| X_trainv, X_valid, Y_trainv, Y_valid = train_test_split(train_x, train_y, test_size=0.2, random_state=16518, shuffle=True) |
|
|
| train_x.shape, valid_x.shape |
|
|
| X_trainv.shape, X_valid.shape |
|
|
| def create_polynomial_regression_model(degree): |
| "Create a polynomial regression model for the given degree" |
|
|
| poly_features = PolynomialFeatures(degree=degree, include_bias=False) |
|
|
| X_train_poly = poly_features.fit_transform(X_trainv) |
|
|
| poly_model = LinearRegression() |
| poly_model.fit(X_train_poly, Y_trainv) |
|
|
| y_train_predicted = poly_model.predict(X_train_poly) |
|
|
| y_valid_predict = poly_model.predict(poly_features.fit_transform(X_valid)) |
|
|
| mse_train = metrics.mean_squared_error(Y_trainv, y_train_predicted) |
|
|
| mse_valid = metrics.mean_squared_error(Y_valid, y_valid_predict) |
|
|
| return (mse_train, mse_valid,degree) |
|
|
| a=[] |
| for i in range(1,8): |
| a.append(create_polynomial_regression_model(i)) |
| df = pd.DataFrame(a,columns=['Train Error', 'Validation Error', 'Degree']) |
| df.sort_values(by='Validation Error') |
|
|
| scaler = StandardScaler() |
| train_x = scaler.fit_transform(train_x) |
| valid_x = scaler.transform(valid_x) |
|
|
| polynomial_features = PolynomialFeatures(degree=2, include_bias=False) |
| train_x_poly = polynomial_features.fit_transform(train_x) |
| valid_x_poly = polynomial_features.fit_transform(valid_x) |
|
|
| polymodel = LinearRegression() |
| polymodel.fit(train_x_poly, train_y) |
| pred = polymodel.predict(valid_x_poly) |
| r2 = metrics.r2_score(valid_y,pred) |
|
|
| print('r2_score:', r2) |