Anusha3 commited on
Commit
d70e5d1
·
verified ·
1 Parent(s): 815ebe7

Upload folder using huggingface_hub

Browse files
backend_files/Dockerfile ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ FROM python:3.9-slim
2
+
3
+ # Set the working directory inside the container
4
+ WORKDIR /app
5
+
6
+ # Copy all files from the current directory to the container's working directory
7
+ COPY . .
8
+
9
+ # Install dependencies from the requirements file without using cache to reduce image size
10
+ RUN pip install --no-cache-dir --upgrade -r requirements.txt
11
+
12
+ # Define the command to start the application using Gunicorn with 4 worker processes
13
+ # - `-w 4`: Uses 4 worker processes for handling requests
14
+ # - `-b 0.0.0.0:7860`: Binds the server to port 7860 on all network interfaces
15
+ # - `app:app`: Runs the Flask app (assuming `app.py` contains the Flask instance named `app`)
16
+ CMD ["gunicorn", "-w", "4", "-b", "0.0.0.0:7860", "app:rental_price_predictor_api"]
backend_files/app.py ADDED
@@ -0,0 +1,94 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ # Import necessary libraries
3
+ import numpy as np
4
+ import joblib # For loading the serialized model
5
+ import pandas as pd # For data manipulation
6
+ from flask import Flask, request, jsonify # For creating the Flask API
7
+
8
+ # Initialize the Flask application
9
+ superkart_sales_api = Flask("SuperKart Sales Predictor")
10
+
11
+ # Load the trained machine learning model
12
+ model = joblib.load("superkart_sales_prediction_model_v1_0.joblib")
13
+
14
+ # Define a route for the home page (GET request)
15
+ @superkart_sales_api.get('/')
16
+ def home():
17
+ """
18
+ This function handles GET requests to the root URL ('/') of the API.
19
+ It returns a simple welcome message.
20
+ """
21
+ return "Welcome to the SuperKart Sales Prediction API!"
22
+
23
+ # Define an endpoint for single sales prediction (POST request)
24
+ @superkart_sales_api.post('/v1/sales')
25
+ def predict_sales():
26
+ """
27
+ This function handles POST requests to the '/v1/sales' endpoint.
28
+ It expects a JSON payload containing product and store details
29
+ and returns the predicted sales as a JSON response.
30
+ """
31
+ # Get the JSON data from the request body
32
+ product_data = request.get_json()
33
+
34
+ # Extract relevant features from the JSON data
35
+ sample = {
36
+ 'Product_Weight': product_data['Product_Weight'],
37
+ 'Product_Allocated_Area': product_data['Product_Allocated_Area'],
38
+ 'Product_MRP': product_data['Product_MRP'],
39
+ 'Store_Establishment_Year': product_data['Store_Establishment_Year'],
40
+ 'Product_Sugar_Content': product_data['Product_Sugar_Content'],
41
+ 'Product_Type': product_data['Product_Type'],
42
+ 'Store_Size': product_data['Store_Size'],
43
+ 'Store_Location_City_Type': product_data['Store_Location_City_Type'],
44
+ 'Store_Type': product_data['Store_Type']
45
+ }
46
+
47
+ # Convert the extracted data into a Pandas DataFrame
48
+ input_data = pd.DataFrame([sample])
49
+
50
+ # Make prediction
51
+ predicted_sales = model.predict(input_data)[0]
52
+
53
+ # Convert predicted_sales to Python float
54
+ predicted_sales = round(float(predicted_sales), 2)
55
+
56
+ # Return the prediction as JSON
57
+ return jsonify({'Predicted Sales (units)': predicted_sales})
58
+
59
+
60
+ # Define an endpoint for batch prediction (POST request)
61
+ @superkart_sales_api.post('/v1/salesbatch')
62
+ def predict_sales_batch():
63
+ """
64
+ This function handles POST requests to the '/v1/salesbatch' endpoint.
65
+ It expects a CSV file containing product details for multiple products
66
+ and returns the predicted sales as a dictionary in the JSON response.
67
+ """
68
+ # Get the uploaded CSV file from the request
69
+ file = request.files['file']
70
+
71
+ # Read the CSV file into a Pandas DataFrame
72
+ input_data = pd.read_csv(file)
73
+
74
+ # Make predictions for all rows in the DataFrame
75
+ predicted_sales = model.predict(input_data).tolist()
76
+
77
+ # Round and convert to float
78
+ predicted_sales = [round(float(sale), 2) for sale in predicted_sales]
79
+
80
+ # If the CSV has an 'id' column, use it as keys; else just index
81
+ if 'id' in input_data.columns:
82
+ ids = input_data['id'].tolist()
83
+ else:
84
+ ids = list(range(1, len(predicted_sales) + 1))
85
+
86
+ output_dict = dict(zip(ids, predicted_sales))
87
+
88
+ # Return the predictions dictionary as a JSON response
89
+ return jsonify(output_dict)
90
+
91
+
92
+ # Run the Flask application in debug mode if this script is executed directly
93
+ if __name__ == '__main__':
94
+ superkart_sales_api.run(debug=True)
backend_files/requirements.txt ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ pandas==2.2.2
2
+ numpy==2.0.2
3
+ scikit-learn==1.6.1
4
+ xgboost==2.1.4
5
+ joblib==1.4.2
6
+ Werkzeug==2.2.2
7
+ flask==2.2.2
8
+ gunicorn==20.1.0
9
+ requests==2.28.1
10
+ uvicorn[standard]
11
+ streamlit==1.43.2
frontend_files/Dockerfile ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Use a minimal base image with Python 3.9 installed
2
+ FROM python:3.9-slim
3
+
4
+ # Set the working directory inside the container to /app
5
+ WORKDIR /app
6
+
7
+ # Copy all files from the current directory on the host to the container's /app directory
8
+ COPY . .
9
+
10
+ # Install Python dependencies listed in requirements.txt
11
+ RUN pip3 install -r requirements.txt
12
+
13
+ # Define the command to run the Streamlit app on port 8501 and make it accessible externally
14
+ CMD ["streamlit", "run", "app.py", "--server.port=8501", "--server.address=0.0.0.0", "--server.enableXsrfProtection=false"]
15
+
16
+ # NOTE: Disable XSRF protection for easier external access in order to make batch predictions
frontend_files/app.py ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ import streamlit as st
3
+ import pandas as pd
4
+ import requests
5
+
6
+ # -------------------------------
7
+ # Streamlit Frontend for Superkart Sales Prediction
8
+ # -------------------------------
9
+
10
+ # Set the title of the Streamlit app
11
+ st.title("Superkart Sales Prediction App")
12
+
13
+ # Section for online prediction
14
+ st.subheader("Online Prediction")
15
+
16
+ # Collect user input for product & store features
17
+ product_weight = st.number_input("Product Weight (grams)", min_value=0.0, value=500.0, step=50.0)
18
+ product_visibility = st.number_input("Product Visibility", min_value=0.0, max_value=1.0, step=0.01, value=0.05)
19
+ product_mrp = st.number_input("Product MRP (₹)", min_value=0.0, value=120.0, step=1.0)
20
+
21
+ outlet_establishment_year = st.number_input("Outlet Establishment Year", min_value=1950, max_value=2030, value=2000, step=1)
22
+ outlet_size = st.selectbox("Outlet Size", ["Small", "Medium", "High"])
23
+ outlet_location_type = st.selectbox("Outlet Location Type", ["Tier 1", "Tier 2", "Tier 3"])
24
+ outlet_type = st.selectbox("Outlet Type", ["Grocery Store", "Supermarket Type1", "Supermarket Type2", "Supermarket Type3"])
25
+
26
+ # Convert user input into a DataFrame
27
+ input_data = pd.DataFrame([{
28
+
29
+ 'Product_Weight': product_weight,
30
+ 'Product_Visibility': product_visibility,
31
+ 'Product_MRP': product_mrp,
32
+ 'Outlet_Establishment_Year': outlet_establishment_year,
33
+ 'Outlet_Size': outlet_size,
34
+ 'Outlet_Location_Type': outlet_location_type,
35
+ 'Outlet_Type': outlet_type
36
+ }])
37
+
38
+ # Make prediction when the "Predict" button is clicked
39
+ if st.button("Predict Sales"):
40
+ response = requests.post(
41
+ "https://<username>-Superkart_Docker_space.hf.space/v1/sales",
42
+ json=input_data.to_dict(orient='records')[0]
43
+ ) # Send data to backend API
44
+
45
+ if response.status_code == 200:
46
+ prediction = response.json()['Predicted Sales']
47
+ st.success(f"Predicted Sales: {prediction}")
48
+ else:
49
+ st.error("Error making prediction. Please check the backend logs.")
50
+
51
+ # Section for batch prediction
52
+ st.subheader("Batch Prediction")
53
+
54
+ # Allow users to upload a CSV file for batch prediction
55
+ uploaded_file = st.file_uploader("Upload CSV file for batch prediction", type=["csv"])
56
+
57
+ # Make batch prediction when the "Predict Batch" button is clicked
58
+ if uploaded_file is not None:
59
+ if st.button("Predict Batch Sales"):
60
+ response = requests.post(
61
+ "https://<username>-Superkart_Docker_space.hf.space/v1/salesbatch",
62
+ files={"file": uploaded_file}
63
+ ) # Send file to backend API
64
+
65
+ if response.status_code == 200:
66
+ predictions = response.json()
67
+ st.success(" Batch predictions completed!")
68
+ st.write(predictions) # Display the predictions
69
+ else:
70
+ st.error("Error making batch prediction.")
frontend_files/requirements.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ pandas==2.2.2
2
+ requests==2.28.1
3
+ streamlit==1.43.2