Dewasheesh commited on
Commit
2045944
·
verified ·
1 Parent(s): ed96c9c

Upload folder using huggingface_hub

Browse files
Dockerfile CHANGED
@@ -1,14 +1,16 @@
1
- # Use a minimal base image with Python 3.9 installed
2
  FROM python:3.9-slim
3
 
4
- # Set the working directory inside the container to /app
5
  WORKDIR /app
6
 
7
- # Copy all files from the current directory on the host to the container's /app directory
8
  COPY . .
9
 
10
- # Install Python dependencies listed in requirements.txt
11
- RUN pip3 install -r requirements.txt
12
 
13
- # Define the command to run the Streamlit app on port 8501 and make it accessible externally
14
- CMD ["streamlit", "run", "app.py", "--server.port=8501", "--server.address=0.0.0.0", "--server.enableXsrfProtection=false"]
 
 
 
 
 
1
  FROM python:3.9-slim
2
 
3
+ # Set the working directory inside the container
4
  WORKDIR /app
5
 
6
+ # Copy all files from the current directory to the container's working directory
7
  COPY . .
8
 
9
+ # Install dependencies from the requirements file without using cache to reduce image size
10
+ RUN pip install --no-cache-dir --upgrade -r requirements.txt
11
 
12
+ # Define the command to start the application using Gunicorn with 4 worker processes
13
+ # - `-w 4`: Uses 4 worker processes for handling requests
14
+ # - `-b 0.0.0.0:7860`: Binds the server to port 7860 on all network interfaces
15
+ # - `app:app`: Runs the Flask app (assuming `app.py` contains the Flask instance named `app`)
16
+ CMD ["gunicorn", "-w", "4", "-b", "0.0.0.0:7860", "app:rental_price_predictor_api"]
app.py CHANGED
@@ -1,52 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
 
2
- import streamlit as st
3
- import pandas as pd
4
- import joblib
5
-
6
- # Load the trained model
7
- def load_model():
8
- return joblib.load("churn_prediction_model_v1_0.joblib")
9
-
10
- model = load_model()
11
-
12
- # Streamlit UI for Customer Churn Prediction
13
- st.title("Customer Churn Prediction App")
14
- st.write("This tool predicts customer churn risk based on their details. Enter the required information below.")
15
-
16
- # Collect user input based on dataset columns
17
- Partner = st.selectbox("Does the customer have a partner?", ["Yes", "No"])
18
- SeniorCitizen = st.selectbox("SeniorCitizen ?", ["Yes", "No"])
19
- Dependents = st.selectbox("Does the customer have dependents?", ["Yes", "No"])
20
- PhoneService = st.selectbox("Does the customer have phone service?", ["Yes", "No"])
21
- InternetService = st.selectbox("Type of Internet Service", ["DSL", "Fiber optic", "No"])
22
- Contract = st.selectbox("Type of Contract", ["Month-to-month", "One year", "Two year"])
23
- PaymentMethod = st.selectbox("Payment Method", ["Electronic check", "Mailed check", "Bank transfer", "Credit card"])
24
- tenure = st.number_input("Tenure (Months with the company)", min_value=0, value=12)
25
- MonthlyCharges = st.number_input("Monthly Charges", min_value=0.0, value=50.0)
26
- TotalCharges = st.number_input("Total Charges", min_value=0.0, value=600.0)
27
-
28
-
29
- # Convert categorical inputs to match model training
30
- input_data = pd.DataFrame([{
31
- 'Partner': Partner,
32
- 'Dependents': Dependents,
33
- 'PhoneService': PhoneService,
34
- 'InternetService': InternetService,
35
- 'Contract': Contract,
36
- 'PaymentMethod': PaymentMethod,
37
- 'tenure': tenure,
38
- 'MonthlyCharges': MonthlyCharges,
39
- 'TotalCharges': TotalCharges,
40
- 'SeniorCitizen': 1 if SeniorCitizen == "Yes" else 0
41
- }])
42
-
43
- # Set classification threshold
44
- classification_threshold = 0.5
45
-
46
- # Predict button
47
- if st.button("Predict"):
48
- prediction_proba = model.predict_proba(input_data)[0, 1]
49
- prediction = (prediction_proba >= classification_threshold).astype(int)
50
- result = "churn" if prediction == 1 else "not churn"
51
- st.write(f"Prediction: The customer is likely to **{result}**.")
52
- st.write(f"Churn Probability: {prediction_proba:.2f}")
 
1
+ # Import necessary libraries
2
+ import numpy as np
3
+ import joblib # For loading the serialized model
4
+ import pandas as pd # For data manipulation
5
+ from flask import Flask, request, jsonify # For creating the Flask API
6
+
7
+ # Initialize the Flask application
8
+ rental_price_predictor_api = Flask("Super Kart Sales Predictor")
9
+
10
+ # Load the trained machine learning model
11
+ model = joblib.load("super_kart_prediction_model_v1_0.joblib")
12
+
13
+ # Define a route for the home page (GET request)
14
+ @rental_price_predictor_api.get('/')
15
+ def home():
16
+ """
17
+ This function handles GET requests to the root URL ('/') of the API.
18
+ It returns a simple welcome message.
19
+ """
20
+ return "Welcome to the Super Kart Sales Predictor API!"
21
+
22
+ # Define an endpoint for single property prediction (POST request)
23
+ @rental_price_predictor_api.post('/v1/superkart')
24
+ def predict_rental_price():
25
+ """
26
+ This function handles POST requests to the '/v1/superkart' endpoint.
27
+ It expects a JSON payload containing property details and returns
28
+ the predicted rental price as a JSON response.
29
+ """
30
+ # Get the JSON data from the request body
31
+ product_data = request.get_json()
32
+
33
+ # Extract relevant features from the JSON data
34
+ sample = {
35
+ 'Product_Weight': product_data['Product_Weight'],
36
+ 'Product_Sugar_Content': product_data['Product_Sugar_Content'],
37
+ 'Product_Allocated_Area': product_data['Product_Allocated_Area'],
38
+ 'Product_Type': product_data['Product_Type'],
39
+ 'Product_MRP': product_data['Product_MRP'],
40
+ 'Store_Id': product_data['Store_Id'],
41
+ 'Store_Establishment_Year': product_data['Store_Establishment_Year'],
42
+ 'Store_Size': product_data['Store_Size'],
43
+ 'Store_Location_City_Type': product_data['Store_Location_City_Type'],
44
+ 'Store_Type': product_data['Store_Type']
45
+ }
46
+
47
+ # Convert the extracted data into a Pandas DataFrame
48
+ input_data = pd.DataFrame([sample])
49
+
50
+ # Make prediction (get log_price)
51
+ predicted_log_price = model.predict(input_data)[0]
52
+
53
+ # Calculate actual price
54
+ predicted_price = np.exp(predicted_log_price)
55
+
56
+ # Convert predicted_price to Python float
57
+ predicted_price = round(float(predicted_price), 2)
58
+ # The conversion above is needed as we convert the model prediction (log price) to actual price using np.exp, which returns predictions as NumPy float32 values.
59
+ # When we send this value directly within a JSON response, Flask's jsonify function encounters a datatype error
60
+
61
+ # Return the actual price
62
+ return jsonify({'Predicted Sales Total': predicted_price})
63
+
64
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
requirements.txt CHANGED
@@ -3,4 +3,9 @@ numpy==2.0.2
3
  scikit-learn==1.6.1
4
  xgboost==2.1.4
5
  joblib==1.4.2
 
 
 
 
 
6
  streamlit==1.43.2
 
3
  scikit-learn==1.6.1
4
  xgboost==2.1.4
5
  joblib==1.4.2
6
+ Werkzeug==2.2.2
7
+ flask==2.2.2
8
+ gunicorn==20.1.0
9
+ requests==2.28.1
10
+ uvicorn[standard]
11
  streamlit==1.43.2
super_kart_prediction_model_v1_0.joblib ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be7fa0ee5de13393a6fc55f1974f724c550a97d299c4a1cf0410a4d68474730e
3
+ size 63812499