jackfroooot commited on
Commit
40a8964
·
verified ·
1 Parent(s): 253a404

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. Dockerfile +10 -9
  2. app.py +60 -95
  3. requirements.txt +1 -5
Dockerfile CHANGED
@@ -1,16 +1,17 @@
 
1
  FROM python:3.9-slim
2
 
3
- # Set the working directory inside the container
4
  WORKDIR /app
5
 
6
- # Copy all files from the current directory to the container's working directory
7
  COPY . .
8
 
9
- # Install dependencies from the requirements file without using cache to reduce image size
10
- RUN pip install --no-cache-dir --upgrade -r requirements.txt
11
 
12
- # Define the command to start the application using Gunicorn with 4 worker processes
13
- # - `-w 4`: Uses 4 worker processes for handling requests
14
- # - `-b 0.0.0.0:7860`: Binds the server to port 7860 on all network interfaces
15
- # - `app:app`: Runs the Flask app (assuming `app.py` contains the Flask instance named `app`)
16
- CMD ["gunicorn", "-w", "4", "-b", "0.0.0.0:7860", "app:cust_predictor_api"]
 
1
+ # Use a minimal base image with Python 3.9 installed
2
  FROM python:3.9-slim
3
 
4
+ # Set the working directory inside the container to /app
5
  WORKDIR /app
6
 
7
+ # Copy all files from the current directory on the host to the container's /app directory
8
  COPY . .
9
 
10
+ # Install Python dependencies listed in requirements.txt
11
+ RUN pip3 install -r requirements.txt
12
 
13
+ # Define the command to run the Streamlit app on port 8501 and make it accessible externally
14
+ #CMD ["streamlit", "run", "app.py", "--server.port=8501", "--server.address=0.0.0.0", "--server.enableXsrfProtection=false"]
15
+ CMD ["streamlit", "run", "app.py", "--server.port=7860", "--server.address=0.0.0.0", "--server.enableXsrfProtection=false"]
16
+
17
+ # NOTE: Disable XSRF protection for easier external access in order to make batch predictions
app.py CHANGED
@@ -1,96 +1,61 @@
1
- # Import necessary libraries
2
- import numpy as np
3
- import joblib # For loading the serialized model
4
- import pandas as pd # For data manipulation
5
- from flask import Flask, request, jsonify # For creating the Flask API
6
 
7
- # Initialize the Flask application
8
- cust_predictor_api = Flask("ExtraaLearn Customer Predictor")
9
-
10
- # Load the trained machine learning model
11
- model = joblib.load("customer_prediction_model_v1_0.joblib")
12
-
13
- # Define a route for the home page (GET request)
14
- @cust_predictor_api.get('/')
15
- def home():
16
- """
17
- This function handles GET requests to the root URL ('/') of the API.
18
- It returns a simple welcome message.
19
- """
20
- return "Welcome to the ExtraaLearn Customer Prediction API!"
21
-
22
- classification_threshold = 0.45
23
-
24
- # Define an endpoint for customer prediction (POST request)
25
- @cust_predictor_api.post('/v1/cust_lead')
26
- def predict_cust_lead():
27
- """
28
- This function handles POST requests to the '/v1/cust_lead' endpoint.
29
- It expects a JSON payload containing customer details and returns
30
- the predicted customer probability as a JSON response.
31
- """
32
- # Get the JSON data from the request body
33
- cust_data = request.get_json()
34
-
35
- # Extract relevant features from the JSON data
36
- sample = {
37
- 'age' : cust_data['age'],
38
- 'current_occupation' : cust_data['current_occupation'],
39
- 'first_interaction' : cust_data['first_interaction'],
40
- 'profile_completed' : cust_data['profile_completed'],
41
- 'website_visits' : cust_data['website_visits'],
42
- 'time_spent_on_website' : cust_data['time_spent_on_website'],
43
- 'page_views_per_visit' : cust_data['page_views_per_visit'],
44
- 'last_activity' : cust_data['last_activity'],
45
- 'print_media_type1' : cust_data['print_media_type1'],
46
- 'print_media_type2' : cust_data['print_media_type2'],
47
- 'digital_media' : cust_data['digital_media'],
48
- 'educational_channels' : cust_data['educational_channels'],
49
- 'referral' : cust_data['referral']
50
- }
51
-
52
- # Convert the extracted data into a Pandas DataFrame
53
- input_data = pd.DataFrame([sample])
54
-
55
- # Make prediction
56
- predicted_cust = model.predict_proba(input_data)[0][1]
57
-
58
- # convert continuous prob as 0/1
59
- predicted_cust = (predicted_cust >= classification_threshold).astype(int)
60
-
61
- # Return the actual prediction status
62
- return jsonify({'Predicted customer status': predicted_cust})
63
-
64
-
65
- # Define an endpoint for batch prediction (POST request)
66
- @cust_predictor_api.post('/v1/cust_lead_batch')
67
- def predict_cust_lead_batch():
68
- """
69
- This function handles POST requests to the '/v1/cust_lead_batch' endpoint.
70
- It expects a CSV file containing property details for multiple properties
71
- and returns the predicted status as a dictionary in the JSON response.
72
- """
73
- # Get the uploaded CSV file from the request
74
- file = request.files['file']
75
-
76
- # Read the CSV file into a Pandas DataFrame
77
- input_data = pd.read_csv(file)
78
-
79
- # Make predictions for all properties in the DataFrame (get log_prices)
80
- predicted_cust_list = model.predict_proba(input_data)[0][1]
81
- predicted_cust_list = predicted_cust_list.tolist()
82
-
83
- # Calculate actual prices
84
- predicted_cust_list = [round(float(np.exp(log_price)), 2) for log_price in predicted_log_prices]
85
- predicted_cust_list = [(predicted_cust >= classification_threshold).astype(int) for predicted_cust in predicted_cust_list]
86
-
87
- # Create a dictionary of predictions with customer IDs as keys
88
- ids = input_data['ID'].tolist()
89
- output_dict = dict(zip(ids, predicted_cust_list))
90
-
91
- # Return the predictions dictionary as a JSON response
92
- return output_dict
93
-
94
- # Run the Flask application in debug mode if this script is executed directly
95
- if __name__ == '__main__':
96
- cust_predictor_api.run(debug=True)
 
 
 
 
 
 
1
 
2
+ import streamlit as st
3
+ import pandas as pd
4
+ import joblib
5
+
6
+ import warnings
7
+ warnings.filterwarnings("ignore", message=".*ScriptRunContext.*")
8
+
9
+ # Load the trained model
10
+ def load_model():
11
+ return joblib.load("customer_prediction_model_v1_0.joblib")
12
+
13
+ model = load_model()
14
+
15
+ # Set the title of the Streamlit app
16
+ st.title("ExtraaLearn Customer Predictor")
17
+ st.subheader("Online Prediction")
18
+
19
+ # Collect user input based on dataset columns
20
+ # Collect user input for property features
21
+ age = st.number_input("age", min_value=5, max_value=90, step=1, value=30)
22
+ website_visits = st.number_input("website_visits", min_value=0, step=1, value=1)
23
+ time_spent_on_website = st.number_input("time_spent_on_website", min_value=0, step=1, value=1)
24
+ page_views_per_visit = st.number_input("page_views_per_visit", min_value=0, step=1, value=1)
25
+ current_occupation = st.selectbox("current_occupation", ["Professional", "Student", "Unemployed"])
26
+ first_interaction = st.selectbox("first_interaction", ["Mobile App", "Website"])
27
+ profile_completed = st.selectbox("profile_completed", ["Medium", "High", "Low"])
28
+ last_activity = st.selectbox("last_activity", ["Website Activity", "Email Activity", "Phone Activity"])
29
+ print_media_type1 = st.selectbox("print_media_type1", ["Yes", "No"])
30
+ print_media_type2 = st.selectbox("print_media_type2", ["Yes", "No"])
31
+ digital_media = st.selectbox("digital_media", ["Yes", "No"])
32
+ educational_channels = st.selectbox("educational_channels", ["Yes", "No"])
33
+ referral = st.selectbox("referral", ["Yes", "No"])
34
+
35
+ # Convert user input into a DataFrame
36
+ input_data = pd.DataFrame([{
37
+ 'age' : 'age',
38
+ 'website_visits' : 'website_visits',
39
+ 'time_spent_on_website' : 'time_spent_on_website',
40
+ 'page_views_per_visit' : 'page_views_per_visit',
41
+ 'current_occupation' : 'current_occupation',
42
+ 'first_interaction' : 'first_interaction',
43
+ 'profile_completed' : 'profile_completed',
44
+ 'last_activity' : 'last_activity',
45
+ 'print_media_type1' : 'print_media_type1',
46
+ 'print_media_type2' : 'print_media_type2',
47
+ 'digital_media' : 'digital_media',
48
+ 'educational_channels' : 'educational_channels',
49
+ 'referral' : 'referral'
50
+ }])
51
+
52
+ # Set classification threshold
53
+ classification_threshold = 0.5
54
+
55
+ # Predict button
56
+ if st.button("Predict"):
57
+ prediction_proba = model.predict_proba(input_data)[0, 1]
58
+ prediction = (prediction_proba >= classification_threshold).astype(int)
59
+ result = "Join" if prediction == 1 else "not join"
60
+ st.write(f"Prediction: The customer is likely to **{result}**.")
61
+ st.write(f"Churn Probability: {prediction_proba:.2f}")
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
requirements.txt CHANGED
@@ -1,11 +1,7 @@
 
1
  pandas==2.2.2
2
  numpy==2.0.2
3
  scikit-learn==1.6.1
4
  xgboost==2.1.4
5
  joblib==1.4.2
6
- Werkzeug==2.2.2
7
- flask==2.2.2
8
- gunicorn==20.1.0
9
- requests==2.28.1
10
- uvicorn[standard]
11
  streamlit==1.43.2
 
1
+
2
  pandas==2.2.2
3
  numpy==2.0.2
4
  scikit-learn==1.6.1
5
  xgboost==2.1.4
6
  joblib==1.4.2
 
 
 
 
 
7
  streamlit==1.43.2