Spaces:
Running
Running
File size: 16,278 Bytes
ea02521 596ce81 3aa2ce4 2a06b1f 6b11277 0bc7df2 273b01b 3aa2ce4 9b92b0d 2a06b1f 273b01b 3aa2ce4 273b01b 2a06b1f 5736887 940de5b 6a7b482 5736887 6a7b482 5736887 6a7b482 940de5b 6a7b482 9b92b0d 273b01b 3aa2ce4 273b01b 6a7b482 ae7dd23 3aa2ce4 ae7dd23 3aa2ce4 ae7dd23 3aa2ce4 ae7dd23 3aa2ce4 ae7dd23 3aa2ce4 273b01b 3aa2ce4 2f66ff4 9b92b0d 3aa2ce4 dce996d 3aa2ce4 273b01b 0933a87 3aa2ce4 aa0696d 3aa2ce4 aa0696d 9b92b0d ae7dd23 273b01b 0933a87 273b01b 3aa2ce4 596ce81 3aa2ce4 aa0696d 596ce81 2a06b1f 3aa2ce4 ae7dd23 3aa2ce4 aa0696d 3aa2ce4 6323c73 2eee636 1d1bb6e ab5da9e d8c258c 9d7c660 b3e0e63 21b61fa 2eee636 596ce81 6323c73 d8c258c 45b95b2 d8c258c 2263afc 0bc7df2 2a06b1f aa0696d 3aa2ce4 aa0696d 2a06b1f 0bc7df2 3aa2ce4 0bc7df2 b5b30fd 3aa2ce4 d57789d 835ddf1 d74e2ab 16f2c1e 21b61fa ca5ee29 c33c795 3aa2ce4 ac1a0c2 3aa2ce4 0922281 2f66ff4 ca5ee29 0bc7df2 9b29685 3aa2ce4 ae7dd23 ca5ee29 3aa2ce4 0238b02 ca5ee29 9b29685 596ce81 3aa2ce4 596ce81 3aa2ce4 23adf11 aa0696d 596ce81 3aa2ce4 aa0696d 3aa2ce4 ca5ee29 9b92b0d ca5ee29 6a7b482 2263afc 470324c 2263afc 6a7b482 ca5ee29 2a06b1f 1708454 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 |
import gradio as gr
from gradio_client import Client, handle_file
from google import genai
import os
from typing import Optional, List, Tuple, Union
from huggingface_hub import whoami
from PIL import Image
from io import BytesIO
import tempfile
import ffmpeg
import json
from datetime import datetime, date
from pathlib import Path
# --- Database Setup ---
DATA_DIR = Path("/data")
DATA_DIR.mkdir(exist_ok=True)
USAGE_DB_PATH = DATA_DIR / "usage_limits.json"
def load_usage_db() -> dict:
"""Load the usage database from disk."""
if USAGE_DB_PATH.exists():
try:
with open(USAGE_DB_PATH, 'r') as f:
return json.load(f)
except Exception as e:
print(f"Error loading usage database: {e}")
return {}
return {}
def save_usage_db(db: dict):
"""Save the usage database to disk."""
try:
with open(USAGE_DB_PATH, 'w') as f:
json.dump(db, f, indent=2)
except Exception as e:
print(f"Error saving usage database: {e}")
def check_and_update_usage(username: str) -> bool:
"""
Check if user has reached daily limit and update usage.
Returns True if user can generate, False if limit reached.
"""
db = load_usage_db()
today = str(date.today())
# Initialize user record if not exists
if username not in db:
db[username] = {"date": today, "count": 0}
user_data = db[username]
# Reset count if it's a new day
if user_data["date"] != today:
user_data["date"] = today
user_data["count"] = 0
# Check if limit reached
if user_data["count"] >= 80:
return False
# Increment count
user_data["count"] += 1
db[username] = user_data
save_usage_db(db)
return True
def get_remaining_generations(username: str) -> int:
"""Get the number of remaining generations for today."""
db = load_usage_db()
today = str(date.today())
if username not in db:
return 80
user_data = db[username]
# Reset if it's a new day
if user_data["date"] != today:
return 80
return max(0, 80 - user_data["count"])
# --- Google Gemini API Configuration ---
GOOGLE_API_KEY = os.getenv("GOOGLE_API_KEY", "")
if not GOOGLE_API_KEY:
raise ValueError("GOOGLE_API_KEY environment variable not set.")
client = genai.Client(api_key=os.environ.get("GOOGLE_API_KEY"))
GEMINI_MODEL_NAME = 'gemini-2.5-flash-image-preview'
def verify_pro_status(token: Optional[Union[gr.OAuthToken, str]]) -> bool:
"""Verifies if the user is a Hugging Face PRO user or part of an enterprise org."""
if not token:
return False
if isinstance(token, gr.OAuthToken):
token_str = token.token
elif isinstance(token, str):
token_str = token
else:
return False
try:
user_info = whoami(token=token_str)
return (
user_info.get("isPro", False) or
any(org.get("isEnterprise", False) for org in user_info.get("orgs", []))
)
except Exception as e:
print(f"Could not verify user's PRO/Enterprise status: {e}")
return False
def get_username(token: Optional[Union[gr.OAuthToken, str]]) -> Optional[str]:
"""Get the username from the token."""
if not token:
return None
if isinstance(token, gr.OAuthToken):
token_str = token.token
elif isinstance(token, str):
token_str = token
else:
return None
try:
user_info = whoami(token=token_str)
username = user_info.get("name", None)
print(f"Username: {username}")
return username
except Exception as e:
print(f"Could not get username: {e}")
return None
def _extract_image_data_from_response(response) -> Optional[bytes]:
"""Helper to extract image data from the model's response."""
if hasattr(response, 'candidates') and response.candidates:
for part in response.candidates[0].content.parts:
if hasattr(part, 'inline_data') and hasattr(part.inline_data, 'data'):
return part.inline_data.data
return None
def _get_video_info(video_path: str) -> Tuple[float, Tuple[int, int]]:
"""Instantly gets the framerate and (width, height) of a video using ffprobe."""
probe = ffmpeg.probe(video_path)
video_stream = next((s for s in probe['streams'] if s['codec_type'] == 'video'), None)
if not video_stream:
raise ValueError("No video stream found in the file.")
framerate = eval(video_stream['avg_frame_rate'])
resolution = (int(video_stream['width']), int(video_stream['height']))
return framerate, resolution
def _resize_image(image_path: str, target_size: Tuple[int, int]) -> str:
"""Resizes an image to a target size and saves it to a new temp file."""
with Image.open(image_path) as img:
if img.size == target_size:
return image_path
resized_img = img.resize(target_size, Image.Resampling.LANCZOS)
suffix = os.path.splitext(image_path)[1] or ".png"
with tempfile.NamedTemporaryFile(delete=False, suffix=suffix) as tmp_file:
resized_img.save(tmp_file.name)
return tmp_file.name
def _trim_first_frame_fast(video_path: str) -> str:
"""Removes exactly the first frame of a video without re-encoding."""
with tempfile.NamedTemporaryFile(delete=False, suffix=".mp4") as tmp_output_file:
output_path = tmp_output_file.name
try:
framerate, _ = _get_video_info(video_path)
if framerate == 0: raise ValueError("Framerate cannot be zero.")
start_time = 1 / framerate
(
ffmpeg
.input(video_path, ss=start_time)
.output(output_path, c='copy', avoid_negative_ts='make_zero')
.run(overwrite_output=True, quiet=True)
)
return output_path
except Exception as e:
raise RuntimeError(f"FFmpeg trim error: {e}")
def _combine_videos_simple(video1_path: str, video2_path: str) -> str:
"""Combines two videos using the fast concat demuxer."""
with tempfile.NamedTemporaryFile(delete=False, mode='w', suffix=".txt") as tmp_list_file:
tmp_list_file.write(f"file '{os.path.abspath(video1_path)}'\n")
tmp_list_file.write(f"file '{os.path.abspath(video2_path)}'\n")
list_file_path = tmp_list_file.name
with tempfile.NamedTemporaryFile(delete=False, suffix=".mp4") as tmp_output_file:
output_path = tmp_output_file.name
try:
(
ffmpeg
.input(list_file_path, format='concat', safe=0)
.output(output_path, c='copy')
.run(overwrite_output=True, quiet=True)
)
return output_path
except ffmpeg.Error as e:
raise RuntimeError(f"FFmpeg combine error: {e.stderr.decode()}")
finally:
if os.path.exists(list_file_path):
os.remove(list_file_path)
def _generate_video_segment(input_image_path: str, output_image_path: str, prompt: str, token: str) -> str:
"""Generates a single video segment using the external service."""
video_client = Client("multimodalart/wan-2-2-first-last-frame", hf_token=token)
result = video_client.predict(
start_image_pil=handle_file(input_image_path),
end_image_pil=handle_file(output_image_path),
prompt=prompt, api_name="/generate_video"
)
return result[0]["video"]
def unified_image_generator(prompt: str, images: Optional[List[str]], previous_video_path: Optional[str], last_frame_path: Optional[str], manual_token: str, oauth_token: Optional[gr.OAuthToken]) -> tuple:
if not (verify_pro_status(oauth_token) or verify_pro_status(manual_token)):
raise gr.Error("Access Denied.")
# Check rate limit
username = get_username(oauth_token) or get_username(manual_token)
if not username:
raise gr.Error("Could not identify user.")
if not check_and_update_usage(username):
raise gr.Error("This demo is made for interactive generations, not automated workflows. You have generated 80 images today, come back tomorrow for more.")
try:
contents = [Image.open(image_path[0]) for image_path in images] if images else []
contents.append(prompt)
response = client.models.generate_content(model=GEMINI_MODEL_NAME, contents=contents)
image_data = _extract_image_data_from_response(response)
if not image_data: raise gr.Error("No image data in response")
with tempfile.NamedTemporaryFile(delete=False, suffix=".png") as tmp:
Image.open(BytesIO(image_data)).save(tmp.name)
output_path = tmp.name
can_create_video = bool(images and len(images) == 1)
can_extend_video = False
if can_create_video and previous_video_path and last_frame_path:
# The crucial check for continuity
if images[0][0] == last_frame_path:
can_extend_video = True
return (output_path, gr.update(visible=can_create_video), gr.update(visible=can_extend_video), gr.update(visible=False))
except Exception as e:
raise gr.Error(f"Image generation failed: {e}. Rephrase your prompt to make image generation explicit and try again")
def create_new_video(input_image_gallery: List[str], prompt_input: str, output_image: str, oauth_token: Optional[gr.OAuthToken]) -> tuple:
if not verify_pro_status(oauth_token): raise gr.Error("Access Denied.")
if not input_image_gallery or not output_image: raise gr.Error("Input/output images required.")
try:
new_segment_path = _generate_video_segment(input_image_gallery[0][0], output_image, prompt_input, oauth_token.token)
return new_segment_path, new_segment_path, output_image
except Exception as e:
raise gr.Error(f"Video creation failed: {e}")
def extend_existing_video(input_image_gallery: List[str], prompt_input: str, output_image: str, previous_video_path: str, oauth_token: Optional[gr.OAuthToken]) -> tuple:
if not verify_pro_status(oauth_token): raise gr.Error("Access Denied.")
if not previous_video_path: raise gr.Error("No previous video to extend.")
if not input_image_gallery or not output_image: raise gr.Error("Input/output images required.")
try:
_, target_resolution = _get_video_info(previous_video_path)
resized_input_path = _resize_image(input_image_gallery[0][0], target_resolution)
resized_output_path = _resize_image(output_image, target_resolution)
new_segment_path = _generate_video_segment(resized_input_path, resized_output_path, prompt_input, oauth_token.token)
trimmed_segment_path = _trim_first_frame_fast(new_segment_path)
final_video_path = _combine_videos_simple(previous_video_path, trimmed_segment_path)
return final_video_path, final_video_path, output_image
except Exception as e:
raise gr.Error(f"Video extension failed: {e}")
css = '''
#sub_title{margin-top: -35px !important}
.tab-wrapper{margin-bottom: -33px !important}
.tabitem{padding: 0px !important}
.fillable{max-width: 980px !important}
.dark .progress-text {color: white}
.logo-dark{display: none}
.dark .logo-dark{display: block !important}
.dark .logo-light{display: none}
.grid-container img{object-fit: contain}
.grid-container {display: grid;grid-template-columns: repeat(2, 1fr)}
.grid-container:has(> .gallery-item:only-child) {grid-template-columns: 1fr}
#wan_ad p{text-align: center;padding: .5em}
'''
with gr.Blocks(theme=gr.themes.Citrus(), css=css) as demo:
gr.HTML('''
<img class="logo-dark" src='https://huggingface.co/spaces/multimodalart/nano-banana/resolve/main/nano_banana_pros.png' style='margin: 0 auto; max-width: 650px' />
<img class="logo-light" src='https://huggingface.co/spaces/multimodalart/nano-banana/resolve/main/nano_banana_pros_light.png' style='margin: 0 auto; max-width: 650px' />
''')
gr.HTML("<h3 style='text-align:center'>Hugging Face PRO users can use Google's Nano Banana (Gemini 2.5 Flash Image Preview) on this Space. <a href='http://huggingface.co/subscribe/pro?source=nana_banana' target='_blank'>Subscribe to PRO</a></h3>", elem_id="sub_title")
pro_message = gr.Markdown(visible=False)
main_interface = gr.Column(visible=False)
previous_video_state = gr.State(None)
last_frame_of_video_state = gr.State(None)
with main_interface:
with gr.Row():
with gr.Column(scale=1):
image_input_gallery = gr.Gallery(label="Upload one or more images here. Leave empty for text-to-image", file_types=["image"], height="auto")
prompt_input = gr.Textbox(label="Prompt", placeholder="Turns this photo into a masterpiece")
generate_button = gr.Button("Generate", variant="primary")
with gr.Column(scale=1):
output_image = gr.Image(label="Output", interactive=False, elem_id="output", type="filepath")
use_image_button = gr.Button("♻️ Use this Image for Next Edit", variant="primary")
with gr.Row():
create_video_button = gr.Button("Create video between the two images 🎥", variant="secondary", visible=False)
extend_video_button = gr.Button("Extend existing video with new scene 🎞️", variant="secondary", visible=False)
with gr.Group(visible=False) as video_group:
video_output = gr.Video(label="Generated Video", show_download_button=True, autoplay=True)
gr.Markdown("Generate more with [Wan 2.2 first-last-frame](https://huggingface.co/spaces/multimodalart/wan-2-2-first-last-frame)", elem_id="wan_ad")
manual_token = gr.Textbox("Manual Token (to use with the API)", visible=False)
gr.Markdown("<h2 style='text-align: center'>Thank you for being a PRO! 🤗</h2>")
login_button = gr.LoginButton()
gr.on(
triggers=[generate_button.click, prompt_input.submit],
fn=unified_image_generator,
inputs=[prompt_input, image_input_gallery, previous_video_state, last_frame_of_video_state, manual_token],
outputs=[output_image, create_video_button, extend_video_button, video_group]
)
use_image_button.click(
fn=lambda img: (
[img] if img else None, None, gr.update(visible=False),
gr.update(visible=False), gr.update(visible=False)
),
inputs=[output_image],
outputs=[image_input_gallery, output_image, create_video_button, extend_video_button, video_group]
)
create_video_button.click(
fn=lambda: gr.update(visible=True), outputs=[video_group]
).then(
fn=create_new_video,
inputs=[image_input_gallery, prompt_input, output_image],
outputs=[video_output, previous_video_state, last_frame_of_video_state],
)
extend_video_button.click(
fn=lambda: gr.update(visible=True), outputs=[video_group]
).then(
fn=extend_existing_video,
inputs=[image_input_gallery, prompt_input, output_image, previous_video_state],
outputs=[video_output, previous_video_state, last_frame_of_video_state],
)
def control_access(profile: Optional[gr.OAuthProfile] = None, oauth_token: Optional[gr.OAuthToken] = None):
if not profile: return gr.update(visible=False), gr.update(visible=False)
if verify_pro_status(oauth_token):
return gr.update(visible=True), gr.update(visible=False)
else:
message = (
"## ✨ Exclusive Access for PRO Users\n\n"
"Thank you for your interest! This app is available exclusively for our Hugging Face **PRO** members.\n\n"
"To unlock this and many other cool stuff, please consider upgrading your account.\n\n"
"### [**Become a PRO Today!**](http://huggingface.co/subscribe/pro?source=nana_banana)"
)
return gr.update(visible=False), gr.update(visible=True, value=message)
demo.load(control_access, inputs=None, outputs=[main_interface, pro_message])
if __name__ == "__main__":
demo.queue(max_size=None, default_concurrency_limit=None).launch(show_error=True) |