NitinBot001 commited on
Commit
fcc60a6
·
verified ·
1 Parent(s): 6faaecb

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -8
app.py CHANGED
@@ -64,19 +64,19 @@ class Config:
64
  EMBEDDING_MODEL = os.getenv("EMBEDDING_MODEL", "text-embedding-ada-002") # Can be changed to compatible embedding model
65
 
66
  # Document Processing
67
- CHUNK_SIZE = 500
68
- CHUNK_OVERLAP = 50
69
 
70
  # Rate Limiting
71
  MAX_RETRIES = 5
72
  RATE_LIMIT_DELAY = 2.0
73
- EMBEDDING_BATCH_SIZE = 10 # OpenAI allows more requests
74
  EMBEDDING_DELAY = 1.0 # Lower delay for OpenAI
75
 
76
  # Model Parameters
77
  TEMPERATURE = 0.5
78
- MAX_OUTPUT_TOKENS = 2000
79
- RETRIEVER_K = 10
80
 
81
  # Paths
82
  INDEX_PATH = "faiss_maize_index"
@@ -86,7 +86,7 @@ config = Config()
86
 
87
  # Request/Response Models
88
  class QueryRequest(BaseModel):
89
- query: str = Field(..., min_length=1, max_length=10000)
90
 
91
  class QueryResponse(BaseModel):
92
  answer: str
@@ -492,8 +492,8 @@ async def process_query(request: QueryRequest):
492
  # Extract sources
493
  sources = []
494
  if 'source_documents' in result:
495
- sources = [doc.page_content[:200] + "..."
496
- for doc in result['source_documents'][:3]]
497
 
498
  # Get token usage
499
  token_usage = {}
 
64
  EMBEDDING_MODEL = os.getenv("EMBEDDING_MODEL", "text-embedding-ada-002") # Can be changed to compatible embedding model
65
 
66
  # Document Processing
67
+ CHUNK_SIZE = 8000
68
+ CHUNK_OVERLAP = 100
69
 
70
  # Rate Limiting
71
  MAX_RETRIES = 5
72
  RATE_LIMIT_DELAY = 2.0
73
+ EMBEDDING_BATCH_SIZE = 20 # OpenAI allows more requests
74
  EMBEDDING_DELAY = 1.0 # Lower delay for OpenAI
75
 
76
  # Model Parameters
77
  TEMPERATURE = 0.5
78
+ MAX_OUTPUT_TOKENS = 200000
79
+ RETRIEVER_K = 20
80
 
81
  # Paths
82
  INDEX_PATH = "faiss_maize_index"
 
86
 
87
  # Request/Response Models
88
  class QueryRequest(BaseModel):
89
+ query: str = Field(..., min_length=1, max_length=100000)
90
 
91
  class QueryResponse(BaseModel):
92
  answer: str
 
492
  # Extract sources
493
  sources = []
494
  if 'source_documents' in result:
495
+ sources = [doc.page_content[:2000] + "..."
496
+ for doc in result['source_documents']]
497
 
498
  # Get token usage
499
  token_usage = {}