Spaces:
Sleeping
Sleeping
File size: 2,383 Bytes
9ee218f 113fa9c 9ee218f 6c5bb1b 9ee218f 6c5bb1b 9ee218f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 |
import joblib
import pandas as pd
from flask import Flask, request, jsonify
# Initialize Flask app with a name
sales_prediction_api = Flask("Customer Churn Predictor")
# Load the trained prediction model
model = joblib.load("sales_prediction_model_v1_0.joblib")
pipeline = joblib.load("sales_prediction_pipeline_v1_0.joblib")
# Define a route for the home page
@sales_prediction_api.get('/')
def home():
return "Welcome to the SuperKart Sales Prediction API!"
# Define an endpoint to predict for a single product
@sales_prediction_api.post('/v1/product')
def predict_sales():
# Get JSON data from the request
product_data = request.get_json()
# Extract relevant features from the input data
sample = {
'Product_Id': product_data['Product_Id'],
'Product_Weight': product_data['Product_Weight'],
'Product_Sugar_Content': product_data['Product_Sugar_Content'],
'Product_Allocated_Area': product_data['Product_Allocated_Area'],
'Product_Type': product_data['Product_Type'],
'Product_MRP': product_data['Product_MRP'],
'Store_Size': product_data['Store_Size'],
'Store_Location_City_Type': product_data['Store_Location_City_Type'],
'Store_Type': product_data['Store_Type']
}
# Convert the extracted data into a DataFrame
input_data = pd.DataFrame([sample])
input_data = pipeline.transform(input_data)
# Make a prediction using the trained model
prediction = model.predict(input_data).tolist()[0]
# Return the prediction as a JSON response
return jsonify({'Prediction': {"Product_Id": product_data['Product_Id'], "Sales": prediction}})
# Define an endpoint to predict sales for a batch of products
@sales_prediction_api.post('/v1/productbatch')
def predict_batch():
# Get the uploaded CSV file from the request
file = request.files['file']
# Read the file into a DataFrame
input_data = pd.read_csv(file)
id_list = input_data.Product_Id.values.tolist()
# Transform the input using the same trained pipeline:
input_data = pipeline.transform(input_data)
# Make predictions for the batch data:
predictions = model.predict(input_data).tolist()
output_dict = dict(zip(id_list, predictions))
return output_dict
# Run the Flask app in debug mode
if __name__ == '__main__':
sales_prediction_api.run(debug=True)
|