KJ24 commited on
Commit
d552e9c
·
verified ·
1 Parent(s): ad15583

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -5
app.py CHANGED
@@ -5,17 +5,17 @@ import torch
5
  import torch.nn.functional as F
6
  import os
7
 
8
- # 📁 Définir le dossier de cache autorisé
9
  CACHE_DIR = "/data"
10
  os.environ["HF_HOME"] = CACHE_DIR
11
  os.environ["TRANSFORMERS_CACHE"] = CACHE_DIR
12
 
13
  app = FastAPI()
14
 
15
- # ✅ Charger le modèle nomic-embed-text-v1
16
  MODEL_NAME = "nomic-ai/nomic-embed-text-v1"
17
- tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, cache_dir=CACHE_DIR)
18
- model = AutoModel.from_pretrained(MODEL_NAME, cache_dir=CACHE_DIR)
19
 
20
  class EmbedInput(BaseModel):
21
  text: str
@@ -25,6 +25,6 @@ async def embed_text(payload: EmbedInput):
25
  inputs = tokenizer(payload.text, return_tensors="pt", padding=True, truncation=True)
26
  with torch.no_grad():
27
  outputs = model(**inputs)
28
- embeddings = outputs.last_hidden_state[:, 0] # CLS token
29
  normalized = F.normalize(embeddings, p=2, dim=1)
30
  return {"embedding": normalized[0].tolist()}
 
5
  import torch.nn.functional as F
6
  import os
7
 
8
+ # 📁 Définir le cache autorisé
9
  CACHE_DIR = "/data"
10
  os.environ["HF_HOME"] = CACHE_DIR
11
  os.environ["TRANSFORMERS_CACHE"] = CACHE_DIR
12
 
13
  app = FastAPI()
14
 
15
+ # ✅ Modèle avec custom code : activer trust_remote_code
16
  MODEL_NAME = "nomic-ai/nomic-embed-text-v1"
17
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, cache_dir=CACHE_DIR, trust_remote_code=True)
18
+ model = AutoModel.from_pretrained(MODEL_NAME, cache_dir=CACHE_DIR, trust_remote_code=True)
19
 
20
  class EmbedInput(BaseModel):
21
  text: str
 
25
  inputs = tokenizer(payload.text, return_tensors="pt", padding=True, truncation=True)
26
  with torch.no_grad():
27
  outputs = model(**inputs)
28
+ embeddings = outputs.last_hidden_state[:, 0]
29
  normalized = F.normalize(embeddings, p=2, dim=1)
30
  return {"embedding": normalized[0].tolist()}