1MR's picture
Update app.py
c45ad33 verified
from fastapi import FastAPI, File, UploadFile
from fastapi.responses import JSONResponse
import tensorflow as tf
import numpy as np
import shutil
import os
from huggingface_hub import InferenceClient
import json
import requests
from langchain_core.output_parsers import StrOutputParser
from langchain_core.runnables import chain
from langchain_huggingface import HuggingFaceEndpoint,ChatHuggingFace
from langchain_core.prompts import ChatPromptTemplate
from langchain_core.runnables import RunnableParallel
from PIL import Image
import json
import time
import requests
from datetime import datetime
# Initialize FastAPI app
app = FastAPI()
chat_nutrition_prompt = ChatPromptTemplate.from_template(
'''
Provide the nutrition information (Calories, Protein, Carbohydrates, Dietary Fiber, Sugars, Fat, Sodium, Potassium, Vitamin C, Vitamin B6) for {prediction} per 100 grams, Output the information as a concise, formatted list without repetition.
'''
)
chat_health_benefits_prompt = ChatPromptTemplate.from_template(
'''
Provide the health benefits and considerations for {prediction}. Additionally, include practical tips for making {prediction} healthier. Keep the response focused on these two aspects only.
'''
)
chat_recipes_prompt = ChatPromptTemplate.from_template(
'''
Tell me about the two most famous recipes for {prediction}. Include the ingredients only.
'''
)
def load_and_prep_image(uploaded_file, img_shape=224):
img = Image.open(uploaded_file) # Open uploaded image
img = img.resize((img_shape, img_shape)) # Resize image
img = tf.convert_to_tensor(img, dtype=tf.float32) # Convert to tensor
return img
@chain
def predict_label(uploaded_file):
img = load_and_prep_image(uploaded_file, img_shape=224) # Preprocess image
img = tf.expand_dims(img, axis=0) # Add batch dimension
pred = model.predict(img) # Model prediction
pred_class_index = np.argmax(pred, axis=1)[0] # Get highest probability index
pred_class_name = class_labels[pred_class_index] # Convert index to class name
return pred_class_name
model = tf.keras.models.load_model("NewVersionModelOptimized40V2.keras")
class_labels = {0: 'Baked Potato',1: 'Burger',2: 'Cake',3: 'Chips',4: 'Crispy Chicken',5: 'Croissant',
6: 'Dount',7: 'Dragon Fruit',8: 'Frise',9: 'Hot Dog',10: 'Jalapeno',11: 'Kiwi',12: 'Lemon',13: 'Lettuce',
14: 'Mango',15: 'Onion',16: 'Orange',17: 'Pizza',18: 'Taquito',19: 'apple',20: 'banana',21: 'beetroot',
22: 'bell pepper',23: 'bread',24: 'cabbage',25: 'carrot',26: 'cauliflower',27: 'cheese',28: 'chilli pepper',
29: 'corn',30: 'crab',31: 'cucumber',32: 'eggplant',33: 'eggs',34: 'garlic',36: 'grapes',37: 'milk',
38: 'salamon',39: 'yogurt'}
api_key='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00'
llm = HuggingFaceEndpoint(
repo_id="Qwen/Qwen2.5-72B-Instruct",
task="text-generation",
max_new_tokens=512,
do_sample=False,
repetition_penalty=1.03,
)
chat= ChatHuggingFace(llm=llm)
str_output_parser = StrOutputParser()
chain_label = predict_label
chain1= chat_nutrition_prompt | chat | str_output_parser
chain2= chat_health_benefits_prompt | chat | str_output_parser
chain3= chat_recipes_prompt | chat | str_output_parser
chain_parallel = RunnableParallel({'chat_nutrition_prompt':chain1,
'chat_health_benefits_prompt':chain2,
'chat_recipes_prompt':chain3})
@app.get("/")
def read_root():
keep_alive()
return {"message": "This is My Nutrionguid App FAST"}
def keep_alive(space_url="https://1mr-apigmail.hf.space/ping", interval_hours=5):
while True:
try:
print(f"🔄 Pinging {space_url} at {datetime.now()}")
response = requests.get(space_url)
if response.status_code == 200:
print("")
else:
print("")
except Exception as e:
print("")
time.sleep(interval_hours * 3600)
@app.post("/predictNUT")
async def predict_image_and_nutrition(file: UploadFile = File(...)):
try:
# Save the uploaded file
file_location = f"./temp_{file.filename}"
with open(file_location, "wb") as f:
shutil.copyfileobj(file.file, f)
# Predict the label using the same prediction logic
with open(file_location, "rb") as image_file:
prediction = predict_label.invoke(image_file)
# Remove the temporary file
# os.remove(file_location)
result = chain_parallel.invoke(prediction)
return {
"Predicted_label": prediction,
"Nutrition_info": result['chat_nutrition_prompt'],
"Information": result['chat_health_benefits_prompt'],
"Recipes":result['chat_recipes_prompt']
}
except Exception as e:
return JSONResponse(
status_code=500,
content={"error": f"An error occurred: {str(e)}"}
)