anithajk commited on
Commit
6e3f95c
·
verified ·
1 Parent(s): a42280f

Upload folder using huggingface_hub

Browse files
Dockerfile CHANGED
@@ -1,16 +1,16 @@
1
- # Use a minimal base image with Python 3.9 installed
2
  FROM python:3.9-slim
3
 
4
- # Set the working directory inside the container to /app
5
  WORKDIR /app
6
 
7
- # Copy all files from the current directory on the host to the container's /app directory
8
  COPY . .
9
 
10
- # Install Python dependencies listed in requirements.txt
11
- RUN pip3 install -r requirements.txt
12
 
13
- # Define the command to run the Streamlit app on port 8501 and make it accessible externally
14
- CMD ["streamlit", "run", "app.py", "--server.port=8501", "--server.address=0.0.0.0", "--server.enableXsrfProtection=false"]
15
-
16
- # NOTE: Disable XSRF protection for easier external access in order to make batch predictions
 
 
 
1
  FROM python:3.9-slim
2
 
3
+ # Set the working directory inside the container
4
  WORKDIR /app
5
 
6
+ # Copy all files from the current directory to the container's working directory
7
  COPY . .
8
 
9
+ # Install dependencies from the requirements file without using cache to reduce image size
10
+ RUN pip install --no-cache-dir --upgrade -r requirements.txt
11
 
12
+ # Define the command to start the application using Gunicorn with 4 worker processes
13
+ # - `-w 4`: Uses 4 worker processes for handling requests
14
+ # - `-b 0.0.0.0:7860`: Binds the server to port 7860 on all network interfaces
15
+ # - `app:app`: Runs the Flask app (assuming `app.py` contains the Flask instance named `app`)
16
+ CMD ["gunicorn", "-w", "4", "-b", "0.0.0.0:7860", "app:superkart_model_api"]
app.py CHANGED
@@ -1,82 +1,95 @@
1
- import streamlit as st
2
- import pandas as pd
3
- import requests
4
-
5
- # Set the title of the Streamlit app
6
- st.title("SuperKart's Deceison Making Model")
7
-
8
- # Section for online prediction
9
- st.subheader("Online SuperKart's Model")
10
-
11
- # Collect user input for property features
12
- # Product features
13
- product_weight = st.number_input("Product Weight (in grams)", min_value=0.0, step=0.1)
14
- product_sugar_content = st.selectbox(
15
- "Product Sugar Content",
16
- ["Low Sugar", "Regular", "No Sugar"] # make sure matches your training dataset categories
17
- )
18
- product_allocated_area = st.number_input(
19
- "Producted Allocated Area (sq. ft.)", min_value=0.01, step=0.01, value=0.01
20
- )
21
- product_type = st.selectbox(
22
- "Product Type",
23
- ["Dairy", "Snack", "Beverage", "Household", "Frozen", "Health"] # adjust based on dataset
24
- )
25
- product_mrp = st.number_input(
26
- "Product MRP (in dollars)", min_value=1.0, step=0.5, value=10.0
27
- )
28
- # Store features
29
- store_establishment_year = st.number_input(
30
- "Store Establishment Year", min_value=1900, max_value=2025, step=1, value=2000
31
- )
32
- store_size = st.selectbox(
33
- "Store Size",
34
- ["Small", "Medium", "High"] # adjust categories to dataset
35
- )
36
- store_location_city_type = st.selectbox(
37
- "Store Location City Type",
38
- ["Tier 1", "Tier 2", "Tier 3"] # typical encoding in retail datasets
39
- )
40
- store_type = st.selectbox(
41
- "Store Type",
42
- ["Grocery Store", "Supermarket", "Hypermarket"] # adjust to match dataset
43
- )
44
-
45
- # Convert user input into a DataFrame
46
- input_data = pd.DataFrame([{
47
- 'product_weight': product_weight,
48
- 'product_sugar_content': product_sugar_content,
49
- 'product_allocated_area': product_allocated_area,
50
- 'product_type': product_type,
51
- 'product_mrp': product_mrp,
52
- 'store_establishment_year': store_establishment_year,
53
- 'store_size': store_size,
54
- 'store_location_city_type': store_location_city_type,
55
- 'store_type':store_type
56
- }])
57
-
58
- # Make prediction when the "Predict" button is clicked
59
- if st.button("Predict"):
60
- response = requests.post("https://anithajk-SuperKart_Model_Deployment.hf.space/v1/productsale", json=input_data.to_dict(orient='records')[0]) # Send data to Flask API
61
- if response.status_code == 200:
62
- prediction = response.json()['Predicted Revenue (in dollars)']
63
- st.success(f"Predicted Revenue (in dollars): {prediction}")
64
- else:
65
- st.error("Error making prediction.")
66
-
67
- # Section for batch prediction
68
- st.subheader("Batch Prediction")
69
-
70
- # Allow users to upload a CSV file for batch prediction
71
- uploaded_file = st.file_uploader("Upload CSV file for batch prediction", type=["csv"])
72
-
73
- # Make batch prediction when the "Predict Batch" button is clicked
74
- if uploaded_file is not None:
75
- if st.button("Predict Batch"):
76
- response = requests.post("https://anithajk-SuperKart_Model_Deployment.hf.space/v1/productsalebatch", files={"file": uploaded_file}) # Send file to Flask API
77
- if response.status_code == 200:
78
- predictions = response.json()
79
- st.success("Batch predictions completed!")
80
- st.write(predictions) # Display the predictions
81
- else:
82
- st.error("Error making batch prediction.")
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Import necessary libraries
2
+ import numpy as np
3
+ import joblib # For loading the serialized model
4
+ import pandas as pd # For data manipulation
5
+ from flask import Flask, request, jsonify # For creating the Flask API
6
+
7
+ # Initialize the Flask application
8
+ superkart_model_api = Flask("SuperKart’s Decision-Making System")
9
+
10
+ # Load the trained machine learning model
11
+ model = joblib.load("deployment_files/superkart_decision_making_model_v1_0.joblib")
12
+
13
+ # Define a route for the home page (GET request)
14
+ @superkart_model_api.get('/')
15
+ def home():
16
+ """
17
+ This function handles GET requests to the root URL ('/') of the API.
18
+ It returns a simple welcome message.
19
+ """
20
+ return "Welcome to the SuperKart’s Decision-Making System API!"
21
+
22
+ # Define an endpoint for single product sale prediction (POST request)
23
+ @superkart_model_api.post('/v1/productsale')
24
+ def predict_product_sales():
25
+ """
26
+ This function handles POST requests to the '/v1/productsale' endpoint.
27
+ It expects a JSON payload containing product and store details and returns
28
+ total revenue by the sale of that particular product in that particular store as a JSON response.
29
+ """
30
+ # Get the JSON data from the request body
31
+ product_data = request.get_json()
32
+
33
+ # Extract relevant features from the JSON data
34
+ sample = {
35
+ 'product_weight': product_data['product_weight'],
36
+ 'product_sugar_content': product_data['product_sugar_content'],
37
+ 'product_allocated_area': product_data['product_allocated_area'],
38
+ 'product_type': product_data['product_type'],
39
+ 'product_mrp': product_data['product_mrp'],
40
+ 'store_establishment_year': product_data['store_establishment_year'],
41
+ 'store_size': product_data['store_size'],
42
+ 'store_location_city_type': product_data['store_location_city_type'],
43
+ 'store_type': product_data['store_type']
44
+ }
45
+
46
+ # Convert the extracted data into a Pandas DataFrame
47
+ input_data = pd.DataFrame([sample])
48
+
49
+ # Make prediction (get log_price)
50
+ predicted_log_price = model.predict(input_data)[0]
51
+ print(f"Predicted log price: {predicted_log_price}")
52
+
53
+ # Calculate actual price
54
+ predicted_price = np.exp(predicted_log_price)
55
+
56
+ # Convert predicted_price to Python float
57
+ predicted_price = round(float(predicted_price), 2)
58
+ # The conversion above is needed as we convert the model prediction (log price) to actual price using np.exp, which returns predictions as NumPy float32 values.
59
+ # When we send this value directly within a JSON response, Flask's jsonify function encounters a datatype error
60
+
61
+ # Return the actual price
62
+ return jsonify({'Total Revenue (in dollars)': predicted_price})
63
+
64
+
65
+ # Define an endpoint for batch prediction (POST request)
66
+ @superkart_model_api.post('/v1/productsalebatch')
67
+ def predict_product_sale_price_batch():
68
+ """
69
+ This function handles POST requests to the '/v1/productsalebatch' endpoint.
70
+ It expects a CSV file containing product and store details and returns the predicted
71
+ total revenue as a dictionary in the JSON response.
72
+
73
+ """
74
+ # Get the uploaded CSV file from the request
75
+ file = request.files['file']
76
+
77
+ # Read the CSV file into a Pandas DataFrame
78
+ input_data = pd.read_csv(file)
79
+
80
+ # Make predictions for all product sale in the stores in the DataFrame (get log_prices)
81
+ predicted_log_prices = model.predict(input_data).tolist()
82
+
83
+ # Calculate actual prices
84
+ predicted_prices = [round(float(np.exp(log_price)), 2) for log_price in predicted_log_prices]
85
+
86
+ # Create a dictionary of predictions with product IDs as keys
87
+ product_ids = input_data['id'].tolist() # Assuming 'id' is the product ID column
88
+ output_dict = dict(zip(product_ids, predicted_prices)) # Use actual prices
89
+
90
+ # Return the predictions dictionary as a JSON response
91
+ return output_dict
92
+
93
+ # Run the Flask application in debug mode if this script is executed directly
94
+ if __name__ == '__main__':
95
+ superkart_model_api.run(debug=True)
requirements.txt CHANGED
@@ -1,3 +1,11 @@
1
  pandas==2.2.2
 
 
 
 
 
 
 
2
  requests==2.28.1
 
3
  streamlit==1.43.2
 
1
  pandas==2.2.2
2
+ numpy==2.0.2
3
+ scikit-learn==1.6.1
4
+ xgboost==2.1.4
5
+ joblib==1.4.2
6
+ Werkzeug==2.2.2
7
+ flask==2.2.2
8
+ gunicorn==20.1.0
9
  requests==2.28.1
10
+ uvicorn[standard]
11
  streamlit==1.43.2
superkart_decision_making_model_v1_0.joblib ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a75f9ddeac467a30178c6147ec5601df2397f3de05d36dae990c142a40b1b2c3
3
+ size 351975