Spaces:
Sleeping
Sleeping
File size: 1,295 Bytes
5b2a441 64f585d 5b2a441 f7d8642 64f585d 5b2a441 64f585d 5b2a441 64f585d 5b2a441 f7d8642 64f585d 5b2a441 f7d8642 64f585d f7d8642 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 |
import torch
from torch import nn
from transformers import AutoImageProcessor, AutoModel
import gradio as gr
import numpy as np
from PIL import Image
# Nome do modelo no Hugging Face Hub
MODEL_NAME = "facebook/dinov2-small"
# Carregando processador e modelo
processor = AutoImageProcessor.from_pretrained(MODEL_NAME)
model = AutoModel.from_pretrained(MODEL_NAME)
# Projeção para 512D (caso a saída seja >512, reduzimos)
projection = nn.Linear(model.config.hidden_size, 512)
def get_embedding(image: Image.Image):
# Preprocessamento
inputs = processor(images=image, return_tensors="pt")
with torch.no_grad():
outputs = model(**inputs)
# Usando o CLS token como embedding da imagem
last_hidden_state = outputs.last_hidden_state # (batch, seq_len, hidden)
embedding = last_hidden_state[:, 0] # pegando o [CLS] token
# Projeta para 512D
embedding_512 = projection(embedding)
# Converte para lista Python
return embedding_512.squeeze().tolist()
# Cria API com Gradio (sem interface visual, apenas endpoint)
iface = gr.Interface(
fn=get_embedding,
inputs=gr.Image(type="pil"),
outputs=gr.JSON(),
live=False,
api_name="embed" # endpoint em /embed
)
if __name__ == "__main__":
iface.launch()
|