Spaces:
Sleeping
Sleeping
shri-jai
commited on
Commit
·
5b9ecc2
1
Parent(s):
fa1cbc5
chore: chatbot,home auth endpoints
Browse files- src/chatbot/embedding.py +0 -4
- src/chatbot/router.py +4 -2
- src/home/router.py +2 -1
src/chatbot/embedding.py
CHANGED
|
@@ -9,11 +9,8 @@ MODEL_ID = "onnx-community/embeddinggemma-300m-ONNX"
|
|
| 9 |
|
| 10 |
class EmbeddingModel:
|
| 11 |
def __init__(self):
|
| 12 |
-
print("Loading tokenizer…")
|
| 13 |
self.tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
|
| 14 |
|
| 15 |
-
print("Downloading ONNX model files…")
|
| 16 |
-
|
| 17 |
self.model_path = hf_hub_download(
|
| 18 |
repo_id=MODEL_ID,
|
| 19 |
filename="onnx/model.onnx"
|
|
@@ -25,7 +22,6 @@ class EmbeddingModel:
|
|
| 25 |
|
| 26 |
model_dir = os.path.dirname(self.model_path)
|
| 27 |
|
| 28 |
-
print("Creating inference session…")
|
| 29 |
self.session = ort.InferenceSession(
|
| 30 |
self.model_path,
|
| 31 |
providers=["CPUExecutionProvider"],
|
|
|
|
| 9 |
|
| 10 |
class EmbeddingModel:
|
| 11 |
def __init__(self):
|
|
|
|
| 12 |
self.tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
|
| 13 |
|
|
|
|
|
|
|
| 14 |
self.model_path = hf_hub_download(
|
| 15 |
repo_id=MODEL_ID,
|
| 16 |
filename="onnx/model.onnx"
|
|
|
|
| 22 |
|
| 23 |
model_dir = os.path.dirname(self.model_path)
|
| 24 |
|
|
|
|
| 25 |
self.session = ort.InferenceSession(
|
| 26 |
self.model_path,
|
| 27 |
providers=["CPUExecutionProvider"],
|
src/chatbot/router.py
CHANGED
|
@@ -2,11 +2,13 @@ import os
|
|
| 2 |
import shutil
|
| 3 |
import tempfile
|
| 4 |
from typing import Optional
|
|
|
|
| 5 |
|
| 6 |
from fastapi import APIRouter, Depends, File, Form, HTTPException, UploadFile
|
| 7 |
from sqlalchemy import text
|
| 8 |
from sqlmodel.ext.asyncio.session import AsyncSession
|
| 9 |
|
|
|
|
| 10 |
from src.core.database import get_async_session
|
| 11 |
from .schemas import ManualTextRequest
|
| 12 |
from .service import store_manual_text
|
|
@@ -23,7 +25,7 @@ from .service import process_pdf_and_store
|
|
| 23 |
router = APIRouter(prefix="/chatbot", tags=["chatbot"])
|
| 24 |
|
| 25 |
@router.post("/tokenize", response_model=TokenizeResponse)
|
| 26 |
-
async def tokenize_text(payload: TokenizeRequest):
|
| 27 |
try:
|
| 28 |
encoded = embedding_model.tokenizer(
|
| 29 |
payload.text,
|
|
@@ -44,7 +46,7 @@ async def tokenize_text(payload: TokenizeRequest):
|
|
| 44 |
|
| 45 |
@router.post("/semantic-search", response_model=list[SemanticSearchResult])
|
| 46 |
async def semantic_search(
|
| 47 |
-
payload: SemanticSearchRequest, session: AsyncSession = Depends(get_async_session)
|
| 48 |
):
|
| 49 |
|
| 50 |
if len(payload.embedding) == 0:
|
|
|
|
| 2 |
import shutil
|
| 3 |
import tempfile
|
| 4 |
from typing import Optional
|
| 5 |
+
from uuid import UUID
|
| 6 |
|
| 7 |
from fastapi import APIRouter, Depends, File, Form, HTTPException, UploadFile
|
| 8 |
from sqlalchemy import text
|
| 9 |
from sqlmodel.ext.asyncio.session import AsyncSession
|
| 10 |
|
| 11 |
+
from src.auth.utils import get_current_user
|
| 12 |
from src.core.database import get_async_session
|
| 13 |
from .schemas import ManualTextRequest
|
| 14 |
from .service import store_manual_text
|
|
|
|
| 25 |
router = APIRouter(prefix="/chatbot", tags=["chatbot"])
|
| 26 |
|
| 27 |
@router.post("/tokenize", response_model=TokenizeResponse)
|
| 28 |
+
async def tokenize_text(payload: TokenizeRequest,user_id: UUID = Depends(get_current_user)):
|
| 29 |
try:
|
| 30 |
encoded = embedding_model.tokenizer(
|
| 31 |
payload.text,
|
|
|
|
| 46 |
|
| 47 |
@router.post("/semantic-search", response_model=list[SemanticSearchResult])
|
| 48 |
async def semantic_search(
|
| 49 |
+
payload: SemanticSearchRequest, session: AsyncSession = Depends(get_async_session), user_id: UUID = Depends(get_current_user)
|
| 50 |
):
|
| 51 |
|
| 52 |
if len(payload.embedding) == 0:
|
src/home/router.py
CHANGED
|
@@ -26,7 +26,8 @@ async def fetch_home_data(
|
|
| 26 |
|
| 27 |
@router.post("/emotion", response_model=BaseResponse[EmotionLogResponse])
|
| 28 |
async def create_or_update_emotion(
|
| 29 |
-
data: EmotionLogCreate, session: AsyncSession = Depends(get_async_session)
|
|
|
|
| 30 |
):
|
| 31 |
record = await add_or_update_emotion(data, session)
|
| 32 |
return {
|
|
|
|
| 26 |
|
| 27 |
@router.post("/emotion", response_model=BaseResponse[EmotionLogResponse])
|
| 28 |
async def create_or_update_emotion(
|
| 29 |
+
data: EmotionLogCreate, session: AsyncSession = Depends(get_async_session),
|
| 30 |
+
user_id: str = Depends(get_current_user),
|
| 31 |
):
|
| 32 |
record = await add_or_update_emotion(data, session)
|
| 33 |
return {
|