KishoreKT commited on
Commit
c5911df
·
verified ·
1 Parent(s): 25c8e96

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. Dockerfile +9 -9
  2. app.py +61 -94
  3. requirements.txt +0 -8
Dockerfile CHANGED
@@ -1,16 +1,16 @@
 
1
  FROM python:3.9-slim
2
 
3
- # Set the working directory inside the container
4
  WORKDIR /app
5
 
6
- # Copy all files from the current directory to the container's working directory
7
  COPY . .
8
 
9
- # Install dependencies from the requirements file without using cache to reduce image size
10
- RUN pip install --no-cache-dir --upgrade -r requirements.txt
11
 
12
- # Define the command to start the application using Gunicorn with 4 worker processes
13
- # - `-w 4`: Uses 4 worker processes for handling requests
14
- # - `-b 0.0.0.0:7860`: Binds the server to port 7860 on all network interfaces
15
- # - `app:app`: Runs the Flask app (assuming `app.py` contains the Flask instance named `app`)
16
- CMD ["gunicorn", "-w", "4", "-b", "0.0.0.0:7860", "app:superkart_predictor_api"]
 
1
+ # Use a minimal base image with Python 3.9 installed
2
  FROM python:3.9-slim
3
 
4
+ # Set the working directory inside the container to /app
5
  WORKDIR /app
6
 
7
+ # Copy all files from the current directory on the host to the container's /app directory
8
  COPY . .
9
 
10
+ # Install Python dependencies listed in requirements.txt
11
+ RUN pip3 install -r requirements.txt
12
 
13
+ # Define the command to run the Streamlit app on port 8501 and make it accessible externally
14
+ CMD ["streamlit", "run", "app.py", "--server.port=8501", "--server.address=0.0.0.0", "--server.enableXsrfProtection=false"]
15
+
16
+ # NOTE: Disable XSRF protection for easier external access in order to make batch predictions
 
app.py CHANGED
@@ -1,94 +1,61 @@
1
- # Import necessary libraries
2
- import numpy as np
3
- import joblib # For loading the serialized model
4
- import pandas as pd # For data manipulation
5
- from flask import Flask, request, jsonify # For creating the Flask API
6
-
7
- # Initialize the Flask application
8
- superkart_predictor_api = Flask("Superkart Predictor Price Predictor")
9
-
10
- # Load the trained machine learning model
11
- model = joblib.load("superKart_prediction_model_v1_0.joblib")
12
-
13
- # Define a route for the home page (GET request)
14
- @superkart_predictor_api.get('/')
15
- def home():
16
- """
17
- This function handles GET requests to the root URL ('/') of the API.
18
- It returns a simple welcome message.
19
- """
20
- return "Welcome to the Superkart_Predictor Price Prediction API!"
21
-
22
- # Define an endpoint for single property prediction (POST request)
23
- @superkart_predictor_api.post('/v1/rental')
24
- def predict_rental_price():
25
- """
26
- This function handles POST requests to the '/v1/rental' endpoint.
27
- It expects a JSON payload containing property details and returns
28
- the predicted rental price as a JSON response.
29
- """
30
- # Get the JSON data from the request body
31
- property_data = request.get_json()
32
-
33
- # Extract relevant features from the JSON data
34
- sample = {
35
- 'Product_Weight': property_data['Product_Weight'],
36
- 'Product_Allocated_Area': property_data['Product_Allocated_Area'],
37
- 'Store_Current_Age': property_data['Store_Current_Age'],
38
- 'Product_MRP': property_data['Product_MRP'],
39
- 'Product_Sugar_Content': property_data['Product_Sugar_Content'],
40
- 'Product_Type': property_data['Product_Type'],
41
- 'Store_Location_City_Type': property_data['Store_Location_City_Type'],
42
- 'Store_Id': property_data['Store_Id'],
43
- 'Store_Type': property_data['Store_Type'],
44
- 'Store_Size': property_data['Store_Size']
45
- }
46
-
47
- # Convert the extracted data into a Pandas DataFrame
48
- input_data = pd.DataFrame([sample])
49
-
50
- # Make prediction (get log_price)
51
- predicted_log_revenue = model.predict(input_data)[0]
52
-
53
- # Calculate actual price
54
- predicted_revenue = np.exp(predicted_log_revenue)
55
-
56
- # Convert predicted_price to Python float
57
- predicted_revenue = round(float(predicted_revenue), 2)
58
- # The conversion above is needed as we convert the model prediction (log price) to actual price using np.exp, which returns predictions as NumPy float32 values.
59
- # When we send this value directly within a JSON response, Flask's jsonify function encounters a datatype error
60
-
61
- # Return the actual price
62
- return jsonify({'Predicted Price (in dollars)': predicted_revenue})
63
-
64
-
65
- # Define an endpoint for batch prediction (POST request)
66
- @superkart_predictor_api.post('/v1/rentalbatch')
67
- def predict_SuperKart_batch():
68
- """
69
- This function handles POST requests to the '/v1/rentalbatch' endpoint.
70
- It expects a CSV file containing property details for multiple properties
71
- and returns the predicted rental prices as a dictionary in the JSON response.
72
- """
73
- # Get the uploaded CSV file from the request
74
- file = request.files['file']
75
-
76
- # Read the CSV file into a Pandas DataFrame
77
- input_data = pd.read_csv(file)
78
-
79
- # Make predictions for all properties in the DataFrame (get log_prices)
80
- predicted_log_revenues = model.predict(input_data).tolist()
81
-
82
- # Calculate actual prices
83
- predicted_revenues = [round(float(np.exp(Product_Store_Sales_Total)), 2) for log_price in predicted_log_revenues]
84
- return predicted_revenues
85
- # Create a dictionary of predictions with property IDs as keys
86
- #property_ids = input_data['id'].tolist() # Assuming 'id' is the property ID column
87
- #output_dict = dict(zip(property_ids, predicted_prices)) # Use actual prices
88
-
89
- # Return the predictions dictionary as a JSON response
90
- #return output_dict
91
-
92
- # Run the Flask application in debug mode if this script is executed directly
93
- if __name__ == '__main__':
94
- superkart_predictor_api.run(debug=True)
 
1
+ import streamlit as st
2
+ import pandas as pd
3
+ import requests
4
+
5
+ # Set the title of the Streamlit app
6
+ st.title("Superkart Revenue Prediction")
7
+
8
+ # Section for online prediction
9
+ st.subheader("Online Prediction")
10
+
11
+ # Collect user input for property features
12
+ product_weight = st.number_input("Product_Weight", min_value=0.0, max_value=1000.0, step=0.1, value=12.66)
13
+ product_sugar_content = st.selectbox("Product_Sugar_Content", ["Low Sugar", "Regular", "No Sugar"])
14
+ product_allocated_area = st.number_input("Product_Allocated_Area", min_value=0.0, max_value=1.0, step=0.001, value=0.027)
15
+ product_type = st.selectbox("Product_Type", ["Frozen Foods", "Dairy", "Canned", "Baking Goods", "Health and Hygiene", "Snack Foods"])
16
+ product_mrp = st.number_input("Product_MRP", min_value=0.0, max_value=1000.0, step=0.1, value=117.08)
17
+ store_id = st.text_input("Store_Id", ["OUT001", "OUT002", "OUT003", "OUT004"])
18
+ store_establishment_year = st.number_input("Store_Establishment_Year", min_value=1900, max_value=2027, step=1, value=2009)
19
+ store_size = st.selectbox("Store_Size", ["Small", "Medium", "High"])
20
+ store_location_city_type = st.selectbox("Store_Location_City_Type", ["Tier 1", "Tier 2", "Tier 3"])
21
+ store_type = st.selectbox("Store_Type", ["Supermarket Type1", "Supermarket Type2", "Departmental Store", "Food Mart"])
22
+ #product_store_sales_total = st.number_input("Product_Store_Sales_Total", min_value=0.0, max_value=10000.0, step=0.1, value=2842.4)
23
+
24
+ input_data = pd.DataFrame([{
25
+ "Product_Weight": product_weight,
26
+ "Product_Sugar_Content": product_sugar_content,
27
+ "Product_Allocated_Area": product_allocated_area,
28
+ "Product_Type": product_type,
29
+ "Product_MRP": product_mrp,
30
+ "Store_Id": store_id,
31
+ "Store_Establishment_Year": store_establishment_year,
32
+ "Store_Size": store_size,
33
+ "Store_Location_City_Type": store_location_city_type,
34
+ "Store_Type": store_type,
35
+ }])
36
+
37
+ # Make prediction when the "Predict" button is clicked
38
+ if st.button("Predict"):
39
+ response = requests.post("https://<username>-<repo_id>.hf.space/v1/revenue", json=input_data.to_dict(orient='records')[0]) # Send data to Flask API
40
+ if response.status_code == 200:
41
+ prediction = response.json()['Predicted Revenue (in dollars)']
42
+ st.success(f"Predicted Rental Revenue (in dollars): {prediction}")
43
+ else:
44
+ st.error("Error making prediction.")
45
+
46
+ # Section for batch prediction
47
+ st.subheader("Batch Prediction")
48
+
49
+ # Allow users to upload a CSV file for batch prediction
50
+ uploaded_file = st.file_uploader("Upload CSV file for batch prediction", type=["csv"])
51
+
52
+ # Make batch prediction when the "Predict Batch" button is clicked
53
+ if uploaded_file is not None:
54
+ if st.button("Predict Batch"):
55
+ response = requests.post("https://<username>-<repo_id>.hf.space/v1/revenuebatch", files={"file": uploaded_file}) # Send file to Flask API
56
+ if response.status_code == 200:
57
+ predictions = response.json()
58
+ st.success("Batch predictions completed!")
59
+ st.write(predictions) # Display the predictions
60
+ else:
61
+ st.error("Error making batch prediction.")
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
requirements.txt CHANGED
@@ -1,11 +1,3 @@
1
  pandas==2.2.2
2
- numpy==2.0.2
3
- scikit-learn==1.6.1
4
- xgboost==2.1.4
5
- joblib==1.4.2
6
- Werkzeug==2.2.2
7
- flask==2.2.2
8
- gunicorn==20.1.0
9
  requests==2.28.1
10
- uvicorn[standard]
11
  streamlit==1.43.2
 
1
  pandas==2.2.2
 
 
 
 
 
 
 
2
  requests==2.28.1
 
3
  streamlit==1.43.2