Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -230,14 +230,12 @@ from fastapi.responses import HTMLResponse, JSONResponse
|
|
| 230 |
from fastapi.staticfiles import StaticFiles
|
| 231 |
from llama_index.core import StorageContext, load_index_from_storage, VectorStoreIndex, SimpleDirectoryReader, ChatPromptTemplate, Settings
|
| 232 |
from llama_index.embeddings.huggingface import HuggingFaceEmbedding
|
|
|
|
| 233 |
from pydantic import BaseModel
|
| 234 |
from fastapi.middleware.cors import CORSMiddleware
|
| 235 |
from fastapi.templating import Jinja2Templates
|
| 236 |
-
from huggingface_hub import InferenceClient
|
| 237 |
-
import json
|
| 238 |
import datetime
|
| 239 |
from simple_salesforce import Salesforce, SalesforceLogin
|
| 240 |
-
from llama_index.llms.huggingface import HuggingFaceLLM
|
| 241 |
|
| 242 |
# Pydantic model for request body
|
| 243 |
class MessageRequest(BaseModel):
|
|
@@ -279,7 +277,7 @@ async def add_security_headers(request: Request, call_next):
|
|
| 279 |
# CORS configuration
|
| 280 |
app.add_middleware(
|
| 281 |
CORSMiddleware,
|
| 282 |
-
allow_origins=["*"], #
|
| 283 |
allow_credentials=True,
|
| 284 |
allow_methods=["*"],
|
| 285 |
allow_headers=["*"],
|
|
@@ -290,13 +288,11 @@ app.mount("/static", StaticFiles(directory="static"), name="static")
|
|
| 290 |
templates = Jinja2Templates(directory="static")
|
| 291 |
|
| 292 |
# LlamaIndex configuration
|
| 293 |
-
Settings.llm =
|
| 294 |
model_name="meta-llama/Meta-Llama-3-8B-Instruct",
|
| 295 |
-
tokenizer_name="meta-llama/Meta-Llama-3-8B-Instruct",
|
| 296 |
-
context_window=3000,
|
| 297 |
token=os.getenv("HF_TOKEN"),
|
| 298 |
max_new_tokens=512,
|
| 299 |
-
|
| 300 |
)
|
| 301 |
|
| 302 |
Settings.embed_model = HuggingFaceEmbedding(
|
|
|
|
| 230 |
from fastapi.staticfiles import StaticFiles
|
| 231 |
from llama_index.core import StorageContext, load_index_from_storage, VectorStoreIndex, SimpleDirectoryReader, ChatPromptTemplate, Settings
|
| 232 |
from llama_index.embeddings.huggingface import HuggingFaceEmbedding
|
| 233 |
+
from llama_index.llms.huggingface_api import HuggingFaceInferenceAPI
|
| 234 |
from pydantic import BaseModel
|
| 235 |
from fastapi.middleware.cors import CORSMiddleware
|
| 236 |
from fastapi.templating import Jinja2Templates
|
|
|
|
|
|
|
| 237 |
import datetime
|
| 238 |
from simple_salesforce import Salesforce, SalesforceLogin
|
|
|
|
| 239 |
|
| 240 |
# Pydantic model for request body
|
| 241 |
class MessageRequest(BaseModel):
|
|
|
|
| 277 |
# CORS configuration
|
| 278 |
app.add_middleware(
|
| 279 |
CORSMiddleware,
|
| 280 |
+
allow_origins=["*"], # Specify allowed origins in production
|
| 281 |
allow_credentials=True,
|
| 282 |
allow_methods=["*"],
|
| 283 |
allow_headers=["*"],
|
|
|
|
| 288 |
templates = Jinja2Templates(directory="static")
|
| 289 |
|
| 290 |
# LlamaIndex configuration
|
| 291 |
+
Settings.llm = HuggingFaceInferenceAPI(
|
| 292 |
model_name="meta-llama/Meta-Llama-3-8B-Instruct",
|
|
|
|
|
|
|
| 293 |
token=os.getenv("HF_TOKEN"),
|
| 294 |
max_new_tokens=512,
|
| 295 |
+
temperature=0.1
|
| 296 |
)
|
| 297 |
|
| 298 |
Settings.embed_model = HuggingFaceEmbedding(
|