Gemma2 / app.py
Dariky's picture
create app.py
6a6492b verified
raw
history blame contribute delete
804 Bytes
import gradio as gr
import torch
from transformers import AutoTokenizer, AutoModel
MODEL_NAME = "BAAI/bge-multilingual-gemma2"
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
model = AutoModel.from_pretrained(MODEL_NAME)
def embed(text):
inputs = tokenizer(
text,
return_tensors="pt",
padding=True,
truncation=True
)
with torch.no_grad():
outputs = model(**inputs)
embeddings = outputs.last_hidden_state[:, 0] # CLS token
return embeddings[0].tolist()
demo = gr.Interface(
fn=embed,
inputs=gr.Textbox(lines=4, placeholder="Enter text in any language"),
outputs="json",
title="BAAI/bge-multilingual-gemma2 Embedding Space",
description="Multilingual embedding model for semantic search & RAG"
)
demo.launch()