Update app.py
Browse files
app.py
CHANGED
|
@@ -3,7 +3,7 @@ import tempfile
|
|
| 3 |
import json
|
| 4 |
import re
|
| 5 |
import gradio as gr
|
| 6 |
-
|
| 7 |
|
| 8 |
print("Google Generative AI SDK version:", genai.__version__)
|
| 9 |
|
|
@@ -48,17 +48,20 @@ def process_image(image, prompt):
|
|
| 48 |
print(f"Received image: {type(image)}")
|
| 49 |
print(f"Received prompt: {prompt}")
|
| 50 |
|
| 51 |
-
api_key =
|
| 52 |
-
|
| 53 |
-
|
| 54 |
-
|
| 55 |
-
|
| 56 |
-
|
| 57 |
-
|
| 58 |
-
|
| 59 |
-
|
| 60 |
-
|
| 61 |
-
|
|
|
|
|
|
|
|
|
|
| 62 |
|
| 63 |
# Create temp file
|
| 64 |
with tempfile.NamedTemporaryFile(delete=False, suffix=".png") as tmp:
|
|
@@ -73,7 +76,8 @@ def process_image(image, prompt):
|
|
| 73 |
print("Temp file:", temp_file)
|
| 74 |
|
| 75 |
# Upload file
|
| 76 |
-
uploaded_file = client.files.upload(file=temp_file)
|
|
|
|
| 77 |
print("Uploaded:", uploaded_file.name)
|
| 78 |
|
| 79 |
# === CREATE MODEL USING genai.GenerativeModel ====
|
|
@@ -83,15 +87,16 @@ def process_image(image, prompt):
|
|
| 83 |
)
|
| 84 |
|
| 85 |
# === GENERATE CONTENT ====
|
| 86 |
-
response = model.generate_content(
|
| 87 |
-
|
| 88 |
-
|
| 89 |
-
|
| 90 |
-
|
| 91 |
-
|
| 92 |
-
|
| 93 |
-
|
| 94 |
-
|
|
|
|
| 95 |
|
| 96 |
print("Raw response:", response.text[:200], "...")
|
| 97 |
|
|
|
|
| 3 |
import json
|
| 4 |
import re
|
| 5 |
import gradio as gr
|
| 6 |
+
import google.generativeai as genai
|
| 7 |
|
| 8 |
print("Google Generative AI SDK version:", genai.__version__)
|
| 9 |
|
|
|
|
| 48 |
print(f"Received image: {type(image)}")
|
| 49 |
print(f"Received prompt: {prompt}")
|
| 50 |
|
| 51 |
+
api_key = random.choice(API_KEYS)
|
| 52 |
+
if not api_key:
|
| 53 |
+
return "ERROR: Missing GOOGLE_API_KEY.", None
|
| 54 |
+
genai.configure(api_key=api_key)
|
| 55 |
+
model_name = INTERNAL_MODEL_MAP.get(model_choice, "gemini-2.5-flash")
|
| 56 |
+
model = genai.GenerativeModel(model_name=model_name,
|
| 57 |
+
generation_config={"temperature": float(0.2), "top_p": float(0.95)})
|
| 58 |
+
|
| 59 |
+
# # Create RAG store
|
| 60 |
+
# RAG_STORE_NAME = "default_rag_store"
|
| 61 |
+
# try:
|
| 62 |
+
# genai.create_rag_store(name=RAG_STORE_NAME)
|
| 63 |
+
# except Exception as e:
|
| 64 |
+
# print("(Info) RAG store exists:", e)
|
| 65 |
|
| 66 |
# Create temp file
|
| 67 |
with tempfile.NamedTemporaryFile(delete=False, suffix=".png") as tmp:
|
|
|
|
| 76 |
print("Temp file:", temp_file)
|
| 77 |
|
| 78 |
# Upload file
|
| 79 |
+
#uploaded_file = client.files.upload(file=temp_file)
|
| 80 |
+
uploaded_file = genai.upload_file(path=temp_file, mime_type="image/jpeg")
|
| 81 |
print("Uploaded:", uploaded_file.name)
|
| 82 |
|
| 83 |
# === CREATE MODEL USING genai.GenerativeModel ====
|
|
|
|
| 87 |
)
|
| 88 |
|
| 89 |
# === GENERATE CONTENT ====
|
| 90 |
+
response = model.generate_content([prompt, uploaded_file])
|
| 91 |
+
# response = model.generate_content(
|
| 92 |
+
# contents=[
|
| 93 |
+
# prompt,
|
| 94 |
+
# uploaded_file # file reference
|
| 95 |
+
# ],
|
| 96 |
+
# request_options={
|
| 97 |
+
# "temperature": 0.2
|
| 98 |
+
# }
|
| 99 |
+
# )
|
| 100 |
|
| 101 |
print("Raw response:", response.text[:200], "...")
|
| 102 |
|