Spaces:
Sleeping
Sleeping
| import joblib | |
| import pandas as pd | |
| from flask import Flask, request, jsonify | |
| # Initialize Flask app with a name | |
| sales_prediction_api = Flask("Customer Churn Predictor") | |
| # Load the trained prediction model | |
| model = joblib.load("sales_prediction_model_v1_0.joblib") | |
| pipeline = joblib.load("sales_prediction_pipeline_v1_0.joblib") | |
| # Define a route for the home page | |
| def home(): | |
| return "Welcome to the SuperKart Sales Prediction API!" | |
| # Define an endpoint to predict for a single product | |
| def predict_sales(): | |
| # Get JSON data from the request | |
| product_data = request.get_json() | |
| # Extract relevant features from the input data | |
| sample = { | |
| 'Product_Id': product_data['Product_Id'], | |
| 'Product_Weight': product_data['Product_Weight'], | |
| 'Product_Sugar_Content': product_data['Product_Sugar_Content'], | |
| 'Product_Allocated_Area': product_data['Product_Allocated_Area'], | |
| 'Product_Type': product_data['Product_Type'], | |
| 'Product_MRP': product_data['Product_MRP'], | |
| 'Store_Size': product_data['Store_Size'], | |
| 'Store_Location_City_Type': product_data['Store_Location_City_Type'], | |
| 'Store_Type': product_data['Store_Type'] | |
| } | |
| # Convert the extracted data into a DataFrame | |
| input_data = pd.DataFrame([sample]) | |
| input_data = pipeline.transform(input_data) | |
| # Make a prediction using the trained model | |
| prediction = model.predict(input_data).tolist()[0] | |
| # Return the prediction as a JSON response | |
| return jsonify({'Prediction': {"Product_Id": product_data['Product_Id'], "Sales": prediction}}) | |
| # Define an endpoint to predict sales for a batch of products | |
| def predict_batch(): | |
| # Get the uploaded CSV file from the request | |
| file = request.files['file'] | |
| # Read the file into a DataFrame | |
| input_data = pd.read_csv(file) | |
| id_list = input_data.Product_Id.values.tolist() | |
| # Transform the input using the same trained pipeline: | |
| input_data = pipeline.transform(input_data) | |
| # Make predictions for the batch data: | |
| predictions = model.predict(input_data).tolist() | |
| output_dict = dict(zip(id_list, predictions)) | |
| return output_dict | |
| # Run the Flask app in debug mode | |
| if __name__ == '__main__': | |
| sales_prediction_api.run(debug=True) | |