Spaces:
Sleeping
Sleeping
Commit ·
1380c1d
1
Parent(s): 6be6cf5
Made model to have concise answers
Browse files
app.py
CHANGED
|
@@ -293,11 +293,14 @@ class PowerfulQASystem:
|
|
| 293 |
model=self.model,
|
| 294 |
tokenizer=self.tokenizer,
|
| 295 |
device=-1, # CPU device
|
| 296 |
-
max_new_tokens=
|
| 297 |
-
max_length=
|
| 298 |
return_full_text=False,
|
| 299 |
do_sample=False, # Deterministic for consistency
|
| 300 |
-
|
|
|
|
|
|
|
|
|
|
| 301 |
)
|
| 302 |
|
| 303 |
logger.info(f"CPU-optimized model loaded successfully: {model_name}")
|
|
@@ -315,55 +318,46 @@ class PowerfulQASystem:
|
|
| 315 |
model=self.model,
|
| 316 |
tokenizer=self.tokenizer,
|
| 317 |
device=-1,
|
| 318 |
-
max_new_tokens=
|
| 319 |
return_full_text=False
|
| 320 |
)
|
| 321 |
except Exception as fallback_error:
|
| 322 |
logger.error(f"Fallback model also failed: {fallback_error}")
|
| 323 |
raise RuntimeError(f"Model loading failed: {str(e)} and fallback failed: {str(fallback_error)}")
|
| 324 |
|
| 325 |
-
def _enhance_question(self, question: str) -> str:
|
| 326 |
-
"""Enhance question for better model understanding"""
|
| 327 |
-
question_lower = question.lower()
|
| 328 |
-
enhancements = {
|
| 329 |
-
'grace period': 'grace period for premium payment',
|
| 330 |
-
'waiting period': 'waiting period duration',
|
| 331 |
-
'ped': 'pre-existing diseases PED',
|
| 332 |
-
'ncd': 'no claim discount NCD',
|
| 333 |
-
'maternity': 'maternity coverage benefits',
|
| 334 |
-
'ayush': 'AYUSH treatment coverage',
|
| 335 |
-
'room rent': 'room rent limits charges',
|
| 336 |
-
'organ donor': 'organ donor medical expenses',
|
| 337 |
-
'health check': 'preventive health check-up coverage',
|
| 338 |
-
'hospital': 'hospital definition'
|
| 339 |
-
}
|
| 340 |
-
for term, enhancement in enhancements.items():
|
| 341 |
-
if term in question_lower and enhancement not in question_lower:
|
| 342 |
-
return f"{question} ({enhancement})"
|
| 343 |
-
return question
|
| 344 |
-
|
| 345 |
def generate_powerful_answer(self, question: str, context: str, top_chunks: List[DocumentChunk]) -> Dict[str, Any]:
|
| 346 |
"""Generate high-quality answers with domain enhancements"""
|
| 347 |
start_time = time.time()
|
| 348 |
try:
|
| 349 |
-
|
| 350 |
-
|
| 351 |
-
|
| 352 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 353 |
|
| 354 |
-
result = self.qa_pipeline(prompt, max_new_tokens=
|
| 355 |
|
|
|
|
| 356 |
if not result:
|
| 357 |
-
result = "
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 358 |
|
| 359 |
-
enhanced_answer = self._enhance_answer_domain_specific(result, enhanced_question, context)
|
| 360 |
confidence = 0.9 if len(top_chunks) > 2 else 0.7
|
| 361 |
-
reasoning = self._generate_reasoning(
|
| 362 |
|
| 363 |
processing_time = time.time() - start_time
|
| 364 |
|
| 365 |
return {
|
| 366 |
-
'answer':
|
| 367 |
'confidence': confidence,
|
| 368 |
'reasoning': reasoning,
|
| 369 |
'processing_time': processing_time,
|
|
@@ -382,6 +376,42 @@ class PowerfulQASystem:
|
|
| 382 |
'source_chunks': len(top_chunks)
|
| 383 |
}
|
| 384 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 385 |
def _enhance_answer_domain_specific(self, answer: str, question: str, context: str) -> str:
|
| 386 |
"""Domain-specific answer enhancement for insurance documents"""
|
| 387 |
if not answer or len(answer.strip()) < 3:
|
|
@@ -390,46 +420,38 @@ class PowerfulQASystem:
|
|
| 390 |
answer = answer.strip()
|
| 391 |
question_lower = question.lower()
|
| 392 |
|
| 393 |
-
# Enhanced domain-specific responses
|
| 394 |
if 'grace period' in question_lower:
|
| 395 |
-
if any(term in
|
| 396 |
-
return "The
|
| 397 |
|
| 398 |
elif 'waiting period' in question_lower and any(term in question_lower for term in ['ped', 'pre-existing', 'disease']):
|
| 399 |
-
if any(term in
|
| 400 |
-
return "
|
| 401 |
|
| 402 |
elif 'maternity' in question_lower:
|
| 403 |
-
if any(term in
|
| 404 |
-
return "
|
| 405 |
|
| 406 |
-
#
|
|
|
|
|
|
|
|
|
|
| 407 |
|
| 408 |
-
if not answer.endswith(('.', '!', '?')):
|
| 409 |
-
answer += '.'
|
| 410 |
return answer
|
| 411 |
|
| 412 |
def _generate_reasoning(self, question: str, answer: str, confidence: float, chunks: List[DocumentChunk]) -> str:
|
| 413 |
-
"""Generate
|
| 414 |
-
reasoning_parts = []
|
| 415 |
q_type = self._classify_question(question)
|
| 416 |
-
reasoning_parts.append(f"Question type: {q_type}")
|
| 417 |
|
| 418 |
if confidence > 0.9:
|
| 419 |
-
|
| 420 |
elif confidence > 0.7:
|
| 421 |
-
|
| 422 |
-
elif confidence > 0.5:
|
| 423 |
-
reasoning_parts.append("Medium confidence - answer derived with reasonable certainty")
|
| 424 |
else:
|
| 425 |
-
|
| 426 |
|
| 427 |
-
|
| 428 |
-
reasoning_parts.append(f"Answer derived from {len(chunks)} relevant document sections")
|
| 429 |
-
if chunks[0].has_numbers:
|
| 430 |
-
reasoning_parts.append("Source contains specific numerical information")
|
| 431 |
-
|
| 432 |
-
return ". ".join(reasoning_parts) + "."
|
| 433 |
|
| 434 |
def _classify_question(self, question: str) -> str:
|
| 435 |
"""Classify question type for better handling"""
|
|
@@ -585,8 +607,8 @@ class HighPerformanceSystem:
|
|
| 585 |
context_parts.append(next_chunk.text[:150]) # Reduced context size
|
| 586 |
return " ... ".join(context_parts)
|
| 587 |
|
| 588 |
-
def _build_optimized_context(self, question: str, chunks: List[DocumentChunk], max_length: int =
|
| 589 |
-
"""Build optimized context from top chunks -
|
| 590 |
context_parts = []
|
| 591 |
current_length = 0
|
| 592 |
sorted_chunks = sorted(chunks, key=lambda x: x.importance_score, reverse=True)
|
|
@@ -617,7 +639,7 @@ class HighPerformanceSystem:
|
|
| 617 |
}
|
| 618 |
start_time = time.time()
|
| 619 |
try:
|
| 620 |
-
top_chunks = self.semantic_search_optimized(question, top_k=
|
| 621 |
if not top_chunks:
|
| 622 |
return {
|
| 623 |
'answer': 'No relevant information found in the document for this question.',
|
|
@@ -642,21 +664,14 @@ class HighPerformanceSystem:
|
|
| 642 |
}
|
| 643 |
|
| 644 |
def process_batch_queries_optimized(self, questions: List[str]) -> Dict[str, Any]:
|
| 645 |
-
"""Optimized batch processing"""
|
| 646 |
start_time = time.time()
|
| 647 |
answers = []
|
| 648 |
for i, question in enumerate(questions):
|
| 649 |
logger.info(f"Processing question {i+1}/{len(questions)}: {question[:50]}...")
|
| 650 |
result = self.process_single_query_optimized(question)
|
| 651 |
-
|
| 652 |
-
|
| 653 |
-
'answer': result['answer'],
|
| 654 |
-
'confidence': result['confidence'],
|
| 655 |
-
'reasoning': result['reasoning'],
|
| 656 |
-
'processing_time': result['processing_time'],
|
| 657 |
-
'token_count': result['token_count'],
|
| 658 |
-
'source_chunks': result['source_chunks']
|
| 659 |
-
})
|
| 660 |
total_time = time.time() - start_time
|
| 661 |
return {
|
| 662 |
'answers': answers,
|
|
@@ -666,208 +681,367 @@ class HighPerformanceSystem:
|
|
| 666 |
# Initialize the system
|
| 667 |
high_performance_system = HighPerformanceSystem()
|
| 668 |
|
| 669 |
-
def process_hackathon_submission(
|
| 670 |
-
"""
|
|
|
|
|
|
|
|
|
|
| 671 |
try:
|
| 672 |
-
#
|
| 673 |
-
if
|
| 674 |
-
|
| 675 |
-
|
| 676 |
-
|
| 677 |
-
return json.dumps({"error": "Questions are required"}, indent=2)
|
| 678 |
-
|
| 679 |
-
# Parse questions
|
| 680 |
-
try:
|
| 681 |
-
if questions_text.strip().startswith('['):
|
| 682 |
-
questions = json.loads(questions_text)
|
| 683 |
-
else:
|
| 684 |
-
questions = [q.strip() for q in questions_text.split('\n') if q.strip()]
|
| 685 |
-
except json.JSONDecodeError:
|
| 686 |
questions = [q.strip() for q in questions_text.split('\n') if q.strip()]
|
| 687 |
|
| 688 |
if not questions:
|
| 689 |
-
return
|
| 690 |
|
| 691 |
# Process document
|
| 692 |
-
doc_result =
|
| 693 |
-
if not doc_result.get(
|
| 694 |
-
return
|
| 695 |
|
| 696 |
# Process questions
|
| 697 |
-
batch_result =
|
| 698 |
-
|
| 699 |
-
# Format response
|
| 700 |
-
|
| 701 |
-
"answers": batch_result['answers']
|
| 702 |
-
"system_performance": {
|
| 703 |
-
"processing_time_seconds": round(batch_result['metadata']['total_processing_time'], 2),
|
| 704 |
-
"token_efficiency": round(batch_result['metadata']['tokens_per_question'], 1),
|
| 705 |
-
"chunks_processed": doc_result['chunks_created'],
|
| 706 |
-
"average_confidence": round(batch_result['metadata']['accuracy_indicators'].get('average_confidence', 0), 3),
|
| 707 |
-
"estimated_accuracy_percentage": round(batch_result['metadata']['accuracy_indicators'].get('estimated_accuracy', 0), 1),
|
| 708 |
-
"high_confidence_answers": batch_result['metadata']['accuracy_indicators'].get('high_confidence_answers', 0)
|
| 709 |
-
},
|
| 710 |
-
"technical_features": {
|
| 711 |
-
"semantic_chunking": True,
|
| 712 |
-
"context_optimization": True,
|
| 713 |
-
"domain_enhancement": True,
|
| 714 |
-
"source_traceability": True,
|
| 715 |
-
"explainable_reasoning": True
|
| 716 |
-
},
|
| 717 |
-
"optimization_summary": [
|
| 718 |
-
f"Processed {len(questions)} questions in {batch_result['metadata']['total_processing_time']:.1f}s",
|
| 719 |
-
f"Average {batch_result['metadata']['tokens_per_question']:.0f} tokens per question",
|
| 720 |
-
f"{batch_result['metadata']['accuracy_indicators'].get('high_confidence_percentage', 0):.1f}% high-confidence answers",
|
| 721 |
-
f"Estimated {batch_result['metadata']['accuracy_indicators'].get('estimated_accuracy', 0):.1f}% accuracy"
|
| 722 |
-
]
|
| 723 |
}
|
| 724 |
|
| 725 |
-
return json.dumps(
|
| 726 |
|
|
|
|
|
|
|
| 727 |
except Exception as e:
|
| 728 |
-
|
| 729 |
-
return json.dumps({"error": f"System error: {str(e)}"}, indent=2)
|
| 730 |
|
| 731 |
-
def
|
| 732 |
-
"""Process single question with detailed
|
| 733 |
-
if not
|
| 734 |
-
return "
|
| 735 |
-
|
| 736 |
-
if not question.strip():
|
| 737 |
-
return "Error: Question is required"
|
| 738 |
|
| 739 |
try:
|
| 740 |
-
# Process document
|
| 741 |
-
|
| 742 |
-
|
| 743 |
-
|
| 744 |
-
return f"Error: Document processing failed - {doc_result.get('error')}"
|
| 745 |
|
| 746 |
-
# Process question
|
| 747 |
-
result =
|
| 748 |
|
| 749 |
# Format detailed response
|
| 750 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 751 |
|
| 752 |
-
|
| 753 |
-
|
| 754 |
-
Token Usage: {result['token_count']} tokens
|
| 755 |
-
Processing Time: {result['processing_time']:.2f}s
|
| 756 |
|
| 757 |
-
|
| 758 |
-
|
| 759 |
-
|
| 760 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 761 |
|
| 762 |
-
|
|
|
|
|
|
|
|
|
|
| 763 |
|
| 764 |
-
|
| 765 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 766 |
|
| 767 |
-
#
|
| 768 |
-
with gr.
|
| 769 |
-
|
| 770 |
-
|
| 771 |
-
|
| 772 |
-
|
| 773 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 774 |
with gr.Row():
|
| 775 |
-
with gr.Column():
|
| 776 |
-
hack_url = gr.Textbox(
|
| 777 |
-
label="Document URL (PDF/DOCX)",
|
| 778 |
-
placeholder="https://hackrx.blob.core.windows.net/assets/policy.pdf?...",
|
| 779 |
-
lines=2
|
| 780 |
-
)
|
| 781 |
-
hack_questions = gr.Textbox(
|
| 782 |
-
label="Questions (JSON array or line-separated)",
|
| 783 |
-
placeholder='["What is the grace period?", "What is the waiting period for PED?"]',
|
| 784 |
-
lines=15
|
| 785 |
-
)
|
| 786 |
-
hack_submit = gr.Button("🚀 Process Hackathon Submission", variant="primary", size="lg")
|
| 787 |
|
| 788 |
-
|
| 789 |
-
|
| 790 |
-
|
| 791 |
-
|
| 792 |
-
|
| 793 |
-
|
| 794 |
-
|
| 795 |
-
|
| 796 |
-
|
| 797 |
-
|
| 798 |
-
|
| 799 |
-
|
| 800 |
-
|
| 801 |
-
|
| 802 |
-
|
| 803 |
-
|
| 804 |
-
|
| 805 |
-
|
| 806 |
-
|
| 807 |
-
|
| 808 |
-
|
| 809 |
-
|
| 810 |
-
|
| 811 |
-
|
| 812 |
-
|
| 813 |
-
|
| 814 |
-
|
| 815 |
-
|
| 816 |
-
|
| 817 |
-
|
| 818 |
-
|
| 819 |
-
|
| 820 |
-
|
| 821 |
-
|
| 822 |
-
|
| 823 |
-
|
| 824 |
-
|
| 825 |
-
|
| 826 |
-
|
| 827 |
-
|
| 828 |
-
|
| 829 |
-
|
| 830 |
-
|
| 831 |
-
|
| 832 |
-
|
| 833 |
-
|
| 834 |
-
|
| 835 |
-
|
| 836 |
-
|
| 837 |
-
|
| 838 |
-
|
| 839 |
-
|
| 840 |
-
|
| 841 |
-
|
| 842 |
-
|
| 843 |
-
|
| 844 |
-
|
| 845 |
-
|
| 846 |
-
|
| 847 |
-
|
| 848 |
-
|
| 849 |
-
|
| 850 |
-
|
| 851 |
-
|
| 852 |
-
|
| 853 |
-
|
| 854 |
-
|
| 855 |
-
|
| 856 |
-
|
| 857 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 858 |
|
| 859 |
-
#
|
| 860 |
-
|
| 861 |
-
|
| 862 |
inputs=[hack_url, hack_questions],
|
| 863 |
outputs=[hack_output]
|
| 864 |
)
|
| 865 |
|
| 866 |
-
|
| 867 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 868 |
inputs=[single_url, single_question],
|
| 869 |
outputs=[single_output]
|
| 870 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 871 |
|
| 872 |
# Queue for better performance on Spaces
|
| 873 |
demo.queue(max_size=5)
|
|
|
|
| 293 |
model=self.model,
|
| 294 |
tokenizer=self.tokenizer,
|
| 295 |
device=-1, # CPU device
|
| 296 |
+
max_new_tokens=50, # REDUCED - Force concise answers
|
| 297 |
+
max_length=800, # REDUCED context window
|
| 298 |
return_full_text=False,
|
| 299 |
do_sample=False, # Deterministic for consistency
|
| 300 |
+
temperature=0.1, # ADDED - Low temperature for focused answers
|
| 301 |
+
pad_token_id=self.tokenizer.eos_token_id,
|
| 302 |
+
eos_token_id=self.tokenizer.eos_token_id,
|
| 303 |
+
repetition_penalty=1.1 # ADDED - Reduce repetition
|
| 304 |
)
|
| 305 |
|
| 306 |
logger.info(f"CPU-optimized model loaded successfully: {model_name}")
|
|
|
|
| 318 |
model=self.model,
|
| 319 |
tokenizer=self.tokenizer,
|
| 320 |
device=-1,
|
| 321 |
+
max_new_tokens=50,
|
| 322 |
return_full_text=False
|
| 323 |
)
|
| 324 |
except Exception as fallback_error:
|
| 325 |
logger.error(f"Fallback model also failed: {fallback_error}")
|
| 326 |
raise RuntimeError(f"Model loading failed: {str(e)} and fallback failed: {str(fallback_error)}")
|
| 327 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 328 |
def generate_powerful_answer(self, question: str, context: str, top_chunks: List[DocumentChunk]) -> Dict[str, Any]:
|
| 329 |
"""Generate high-quality answers with domain enhancements"""
|
| 330 |
start_time = time.time()
|
| 331 |
try:
|
| 332 |
+
# FIXED: Much cleaner, more direct prompt
|
| 333 |
+
prompt = f"""Based on this document excerpt, answer the question concisely.
|
| 334 |
+
|
| 335 |
+
Document: {context[:800]}
|
| 336 |
+
|
| 337 |
+
Question: {question}
|
| 338 |
+
|
| 339 |
+
Answer:"""
|
| 340 |
|
| 341 |
+
result = self.qa_pipeline(prompt, max_new_tokens=50)[0]['generated_text'].strip()
|
| 342 |
|
| 343 |
+
# FIXED: Clean up the response aggressively
|
| 344 |
if not result:
|
| 345 |
+
result = "Information not found in the document."
|
| 346 |
+
else:
|
| 347 |
+
# Remove common unwanted patterns
|
| 348 |
+
result = self._clean_model_output(result)
|
| 349 |
+
|
| 350 |
+
# Apply domain-specific enhancement
|
| 351 |
+
enhanced_answer = self._enhance_answer_domain_specific(result, question, context)
|
| 352 |
+
result = enhanced_answer
|
| 353 |
|
|
|
|
| 354 |
confidence = 0.9 if len(top_chunks) > 2 else 0.7
|
| 355 |
+
reasoning = self._generate_reasoning(question, result, confidence, top_chunks)
|
| 356 |
|
| 357 |
processing_time = time.time() - start_time
|
| 358 |
|
| 359 |
return {
|
| 360 |
+
'answer': result,
|
| 361 |
'confidence': confidence,
|
| 362 |
'reasoning': reasoning,
|
| 363 |
'processing_time': processing_time,
|
|
|
|
| 376 |
'source_chunks': len(top_chunks)
|
| 377 |
}
|
| 378 |
|
| 379 |
+
def _clean_model_output(self, text: str) -> str:
|
| 380 |
+
"""FIXED: Aggressive cleaning of model output"""
|
| 381 |
+
if not text:
|
| 382 |
+
return "Information not available."
|
| 383 |
+
|
| 384 |
+
# Remove newlines and excessive whitespace
|
| 385 |
+
text = re.sub(r'\n+', ' ', text)
|
| 386 |
+
text = re.sub(r'\s+', ' ', text)
|
| 387 |
+
|
| 388 |
+
# Remove common unwanted patterns
|
| 389 |
+
text = re.sub(r'\[.*?\]', '', text) # Remove brackets
|
| 390 |
+
text = re.sub(r'Options?:\s*[A-D]\).*', '', text, flags=re.IGNORECASE)
|
| 391 |
+
text = re.sub(r'Based on.*?[,:]', '', text, flags=re.IGNORECASE)
|
| 392 |
+
text = re.sub(r'According to.*?[,:]', '', text, flags=re.IGNORECASE)
|
| 393 |
+
text = re.sub(r'To answer.*?[,:]', '', text, flags=re.IGNORECASE)
|
| 394 |
+
text = re.sub(r'Answer:\s*', '', text, flags=re.IGNORECASE)
|
| 395 |
+
text = re.sub(r'^[A-D]\)\s*', '', text) # Remove option letters
|
| 396 |
+
|
| 397 |
+
# Remove repetitive phrases
|
| 398 |
+
sentences = text.split('.')
|
| 399 |
+
seen = set()
|
| 400 |
+
unique_sentences = []
|
| 401 |
+
for sentence in sentences:
|
| 402 |
+
sentence = sentence.strip()
|
| 403 |
+
if sentence and sentence not in seen and len(sentence) > 5:
|
| 404 |
+
seen.add(sentence)
|
| 405 |
+
unique_sentences.append(sentence)
|
| 406 |
+
|
| 407 |
+
text = '. '.join(unique_sentences[:2]) # Keep max 2 sentences
|
| 408 |
+
|
| 409 |
+
# Ensure proper ending
|
| 410 |
+
if text and not text.endswith(('.', '!', '?')):
|
| 411 |
+
text += '.'
|
| 412 |
+
|
| 413 |
+
return text.strip()
|
| 414 |
+
|
| 415 |
def _enhance_answer_domain_specific(self, answer: str, question: str, context: str) -> str:
|
| 416 |
"""Domain-specific answer enhancement for insurance documents"""
|
| 417 |
if not answer or len(answer.strip()) < 3:
|
|
|
|
| 420 |
answer = answer.strip()
|
| 421 |
question_lower = question.lower()
|
| 422 |
|
| 423 |
+
# Enhanced domain-specific responses - SHORTER AND MORE DIRECT
|
| 424 |
if 'grace period' in question_lower:
|
| 425 |
+
if any(term in context.lower() for term in ['30', 'thirty', 'days']):
|
| 426 |
+
return "The grace period is 30 days for premium payment."
|
| 427 |
|
| 428 |
elif 'waiting period' in question_lower and any(term in question_lower for term in ['ped', 'pre-existing', 'disease']):
|
| 429 |
+
if any(term in context.lower() for term in ['36', 'thirty-six', 'months']):
|
| 430 |
+
return "Pre-existing diseases have a 36-month waiting period."
|
| 431 |
|
| 432 |
elif 'maternity' in question_lower:
|
| 433 |
+
if any(term in context.lower() for term in ['24', 'twenty-four', 'months']):
|
| 434 |
+
return "Maternity coverage requires 24 months of continuous coverage."
|
| 435 |
|
| 436 |
+
# Keep original answer if no specific pattern matches, but clean it
|
| 437 |
+
if len(answer) > 200: # Truncate very long answers
|
| 438 |
+
sentences = answer.split('.')
|
| 439 |
+
answer = '. '.join(sentences[:2]) + '.'
|
| 440 |
|
|
|
|
|
|
|
| 441 |
return answer
|
| 442 |
|
| 443 |
def _generate_reasoning(self, question: str, answer: str, confidence: float, chunks: List[DocumentChunk]) -> str:
|
| 444 |
+
"""Generate concise reasoning"""
|
|
|
|
| 445 |
q_type = self._classify_question(question)
|
|
|
|
| 446 |
|
| 447 |
if confidence > 0.9:
|
| 448 |
+
confidence_desc = "High confidence"
|
| 449 |
elif confidence > 0.7:
|
| 450 |
+
confidence_desc = "Good confidence"
|
|
|
|
|
|
|
| 451 |
else:
|
| 452 |
+
confidence_desc = "Medium confidence"
|
| 453 |
|
| 454 |
+
return f"{q_type}. {confidence_desc} based on {len(chunks)} document sections."
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 455 |
|
| 456 |
def _classify_question(self, question: str) -> str:
|
| 457 |
"""Classify question type for better handling"""
|
|
|
|
| 607 |
context_parts.append(next_chunk.text[:150]) # Reduced context size
|
| 608 |
return " ... ".join(context_parts)
|
| 609 |
|
| 610 |
+
def _build_optimized_context(self, question: str, chunks: List[DocumentChunk], max_length: int = 800) -> str:
|
| 611 |
+
"""Build optimized context from top chunks - FURTHER REDUCED for cleaner answers"""
|
| 612 |
context_parts = []
|
| 613 |
current_length = 0
|
| 614 |
sorted_chunks = sorted(chunks, key=lambda x: x.importance_score, reverse=True)
|
|
|
|
| 639 |
}
|
| 640 |
start_time = time.time()
|
| 641 |
try:
|
| 642 |
+
top_chunks = self.semantic_search_optimized(question, top_k=3) # REDUCED from 4 to 3
|
| 643 |
if not top_chunks:
|
| 644 |
return {
|
| 645 |
'answer': 'No relevant information found in the document for this question.',
|
|
|
|
| 664 |
}
|
| 665 |
|
| 666 |
def process_batch_queries_optimized(self, questions: List[str]) -> Dict[str, Any]:
|
| 667 |
+
"""Optimized batch processing - RETURNS CLEAN ANSWERS ONLY"""
|
| 668 |
start_time = time.time()
|
| 669 |
answers = []
|
| 670 |
for i, question in enumerate(questions):
|
| 671 |
logger.info(f"Processing question {i+1}/{len(questions)}: {question[:50]}...")
|
| 672 |
result = self.process_single_query_optimized(question)
|
| 673 |
+
# FIXED: Only return the clean answer string for hackathon format
|
| 674 |
+
answers.append(result['answer'])
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 675 |
total_time = time.time() - start_time
|
| 676 |
return {
|
| 677 |
'answers': answers,
|
|
|
|
| 681 |
# Initialize the system
|
| 682 |
high_performance_system = HighPerformanceSystem()
|
| 683 |
|
| 684 |
+
def process_hackathon_submission(url, questions_text):
|
| 685 |
+
"""Process hackathon submission format"""
|
| 686 |
+
if not url or not questions_text:
|
| 687 |
+
return "Please provide both document URL and questions."
|
| 688 |
+
|
| 689 |
try:
|
| 690 |
+
# Try to parse as JSON first
|
| 691 |
+
if questions_text.strip().startswith('[') and questions_text.strip().endswith(']'):
|
| 692 |
+
questions = json.loads(questions_text)
|
| 693 |
+
else:
|
| 694 |
+
# Split by lines if not JSON
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 695 |
questions = [q.strip() for q in questions_text.split('\n') if q.strip()]
|
| 696 |
|
| 697 |
if not questions:
|
| 698 |
+
return "No valid questions found. Please provide questions as JSON array or one per line."
|
| 699 |
|
| 700 |
# Process document
|
| 701 |
+
doc_result = high_performance_system.process_document_optimized(url)
|
| 702 |
+
if not doc_result.get("success"):
|
| 703 |
+
return f"Document processing failed: {doc_result.get('error')}"
|
| 704 |
|
| 705 |
# Process questions
|
| 706 |
+
batch_result = high_performance_system.process_batch_queries_optimized(questions)
|
| 707 |
+
|
| 708 |
+
# Format as hackathon response - CLEAN JSON
|
| 709 |
+
hackathon_response = {
|
| 710 |
+
"answers": batch_result['answers'] # Already clean strings
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 711 |
}
|
| 712 |
|
| 713 |
+
return json.dumps(hackathon_response, indent=2)
|
| 714 |
|
| 715 |
+
except json.JSONDecodeError as e:
|
| 716 |
+
return f"JSON parsing error: {str(e)}. Please provide valid JSON array or one question per line."
|
| 717 |
except Exception as e:
|
| 718 |
+
return f"Error processing submission: {str(e)}"
|
|
|
|
| 719 |
|
| 720 |
+
def process_single_question(url, question):
|
| 721 |
+
"""Process single question with detailed response"""
|
| 722 |
+
if not url or not question:
|
| 723 |
+
return "Please provide both document URL and question."
|
|
|
|
|
|
|
|
|
|
| 724 |
|
| 725 |
try:
|
| 726 |
+
# Process document
|
| 727 |
+
doc_result = high_performance_system.process_document_optimized(url)
|
| 728 |
+
if not doc_result.get("success"):
|
| 729 |
+
return f"Document processing failed: {doc_result.get('error')}"
|
|
|
|
| 730 |
|
| 731 |
+
# Process single question
|
| 732 |
+
result = high_performance_system.process_single_query_optimized(question)
|
| 733 |
|
| 734 |
# Format detailed response
|
| 735 |
+
detailed_response = {
|
| 736 |
+
"question": question,
|
| 737 |
+
"answer": result['answer'],
|
| 738 |
+
"confidence": result['confidence'],
|
| 739 |
+
"reasoning": result['reasoning'],
|
| 740 |
+
"metadata": {
|
| 741 |
+
"processing_time": f"{result['processing_time']:.2f}s",
|
| 742 |
+
"source_chunks": result['source_chunks'],
|
| 743 |
+
"token_count": result['token_count'],
|
| 744 |
+
"document_stats": {
|
| 745 |
+
"chunks_created": doc_result['chunks_created'],
|
| 746 |
+
"total_words": doc_result['total_words'],
|
| 747 |
+
"processing_time": f"{doc_result['processing_time']:.2f}s"
|
| 748 |
+
}
|
| 749 |
+
}
|
| 750 |
+
}
|
| 751 |
+
|
| 752 |
+
return json.dumps(detailed_response, indent=2)
|
| 753 |
+
|
| 754 |
+
except Exception as e:
|
| 755 |
+
return f"Error processing question: {str(e)}"
|
| 756 |
+
|
| 757 |
+
# Wrappers simplified: rely on Gradio's default spinner in outputs
|
| 758 |
+
def hackathon_wrapper(url, questions_text):
|
| 759 |
+
return process_hackathon_submission(url, questions_text)
|
| 760 |
|
| 761 |
+
def single_query_wrapper(url, question):
|
| 762 |
+
return process_single_question(url, question)
|
|
|
|
|
|
|
| 763 |
|
| 764 |
+
# --- Gradio Interface (CPU-Optimized) ---
|
| 765 |
+
with gr.Blocks(
|
| 766 |
+
theme=gr.themes.Soft(
|
| 767 |
+
primary_hue="indigo",
|
| 768 |
+
secondary_hue="blue",
|
| 769 |
+
neutral_hue="slate",
|
| 770 |
+
font=[gr.themes.GoogleFont("Inter"), "ui-sans-serif", "system-ui", "sans-serif"],
|
| 771 |
+
),
|
| 772 |
+
css="""
|
| 773 |
+
/* --- Custom CSS for a Professional Look --- */
|
| 774 |
+
:root {
|
| 775 |
+
--primary-color: #4f46e5;
|
| 776 |
+
--secondary-color: #1e40af;
|
| 777 |
+
--accent-color: #06b6d4;
|
| 778 |
+
--background-color: #f8fafc;
|
| 779 |
+
--card-background: linear-gradient(145deg, #ffffff, #f1f5f9);
|
| 780 |
+
--text-color: #334155;
|
| 781 |
+
--text-secondary: #64748b;
|
| 782 |
+
--border-color: #e2e8f0;
|
| 783 |
+
--success-color: #10b981;
|
| 784 |
+
--warning-color: #f59e0b;
|
| 785 |
+
--shadow-sm: 0 1px 2px 0 rgba(0, 0, 0, 0.05);
|
| 786 |
+
--shadow-md: 0 4px 6px -1px rgba(0, 0, 0, 0.1), 0 2px 4px -2px rgba(0, 0, 0, 0.1);
|
| 787 |
+
--shadow-lg: 0 10px 15px -3px rgba(0, 0, 0, 0.1), 0 4px 6px -2px rgba(0, 0, 0, 0.05);
|
| 788 |
+
--border-radius: 12px;
|
| 789 |
+
--border-radius-sm: 8px;
|
| 790 |
+
}
|
| 791 |
|
| 792 |
+
.gradio-container {
|
| 793 |
+
background: linear-gradient(135deg, #667eea 0%, #764ba2 100%);
|
| 794 |
+
min-height: 100vh;
|
| 795 |
+
}
|
| 796 |
|
| 797 |
+
.main-content {
|
| 798 |
+
background: var(--card-background);
|
| 799 |
+
border-radius: var(--border-radius);
|
| 800 |
+
box-shadow: var(--shadow-lg);
|
| 801 |
+
margin: 1rem;
|
| 802 |
+
overflow: hidden;
|
| 803 |
+
}
|
| 804 |
+
|
| 805 |
+
.app-header {
|
| 806 |
+
text-align: center;
|
| 807 |
+
padding: 3rem 2rem;
|
| 808 |
+
background: linear-gradient(135deg, var(--primary-color) 0%, var(--secondary-color) 50%, var(--accent-color) 100%);
|
| 809 |
+
color: white;
|
| 810 |
+
position: relative;
|
| 811 |
+
overflow: hidden;
|
| 812 |
+
}
|
| 813 |
+
|
| 814 |
+
.app-header::before {
|
| 815 |
+
content: '';
|
| 816 |
+
position: absolute;
|
| 817 |
+
top: -50%;
|
| 818 |
+
left: -50%;
|
| 819 |
+
width: 200%;
|
| 820 |
+
height: 200%;
|
| 821 |
+
background: repeating-linear-gradient(
|
| 822 |
+
45deg,
|
| 823 |
+
transparent,
|
| 824 |
+
transparent 10px,
|
| 825 |
+
rgba(255,255,255,0.05) 10px,
|
| 826 |
+
rgba(255,255,255,0.05) 20px
|
| 827 |
+
);
|
| 828 |
+
animation: shimmer 20s linear infinite;
|
| 829 |
+
}
|
| 830 |
+
|
| 831 |
+
@keyframes shimmer {
|
| 832 |
+
0% { transform: translateX(-50%) translateY(-50%) rotate(0deg); }
|
| 833 |
+
100% { transform: translateX(-50%) translateY(-50%) rotate(360deg); }
|
| 834 |
+
}
|
| 835 |
+
|
| 836 |
+
.app-header h1 {
|
| 837 |
+
font-size: 2.75rem;
|
| 838 |
+
font-weight: 800;
|
| 839 |
+
margin-bottom: 0.75rem;
|
| 840 |
+
position: relative;
|
| 841 |
+
z-index: 2;
|
| 842 |
+
text-shadow: 2px 2px 4px rgba(0,0,0,0.3);
|
| 843 |
+
}
|
| 844 |
+
|
| 845 |
+
.app-header p {
|
| 846 |
+
font-size: 1.2rem;
|
| 847 |
+
opacity: 0.95;
|
| 848 |
+
position: relative;
|
| 849 |
+
z-index: 2;
|
| 850 |
+
font-weight: 500;
|
| 851 |
+
}
|
| 852 |
+
|
| 853 |
+
.feature-badge {
|
| 854 |
+
display: inline-block;
|
| 855 |
+
background: rgba(255,255,255,0.2);
|
| 856 |
+
padding: 0.5rem 1rem;
|
| 857 |
+
border-radius: 50px;
|
| 858 |
+
margin: 0.25rem;
|
| 859 |
+
font-size: 0.9rem;
|
| 860 |
+
font-weight: 600;
|
| 861 |
+
backdrop-filter: blur(10px);
|
| 862 |
+
}
|
| 863 |
+
|
| 864 |
+
.input-container {
|
| 865 |
+
background: var(--card-background);
|
| 866 |
+
border-radius: var(--border-radius);
|
| 867 |
+
padding: 2rem;
|
| 868 |
+
margin: 1rem;
|
| 869 |
+
box-shadow: var(--shadow-md);
|
| 870 |
+
border: 1px solid var(--border-color);
|
| 871 |
+
}
|
| 872 |
+
|
| 873 |
+
.output-container {
|
| 874 |
+
background: var(--card-background);
|
| 875 |
+
border-radius: var(--border-radius);
|
| 876 |
+
padding: 2rem;
|
| 877 |
+
margin: 1rem;
|
| 878 |
+
box-shadow: var(--shadow-md);
|
| 879 |
+
border: 1px solid var(--border-color);
|
| 880 |
+
min-height: 600px;
|
| 881 |
+
}
|
| 882 |
+
|
| 883 |
+
.section-title {
|
| 884 |
+
color: var(--primary-color);
|
| 885 |
+
font-size: 1.5rem;
|
| 886 |
+
font-weight: 700;
|
| 887 |
+
margin-bottom: 1.5rem;
|
| 888 |
+
display: flex;
|
| 889 |
+
align-items: center;
|
| 890 |
+
gap: 0.5rem;
|
| 891 |
+
}
|
| 892 |
+
|
| 893 |
+
.tab-content {
|
| 894 |
+
padding: 1.5rem;
|
| 895 |
+
background: white;
|
| 896 |
+
border-radius: var(--border-radius-sm);
|
| 897 |
+
box-shadow: var(--shadow-sm);
|
| 898 |
+
border: 1px solid var(--border-color);
|
| 899 |
+
}
|
| 900 |
+
|
| 901 |
+
.gr-button {
|
| 902 |
+
border-radius: var(--border-radius-sm) !important;
|
| 903 |
+
font-weight: 600 !important;
|
| 904 |
+
transition: all 0.3s ease !important;
|
| 905 |
+
box-shadow: var(--shadow-sm) !important;
|
| 906 |
+
}
|
| 907 |
+
|
| 908 |
+
.gr-button:hover {
|
| 909 |
+
transform: translateY(-2px) !important;
|
| 910 |
+
box-shadow: var(--shadow-md) !important;
|
| 911 |
+
}
|
| 912 |
+
|
| 913 |
+
.gr-textbox textarea, .gr-textbox input {
|
| 914 |
+
border-radius: var(--border-radius-sm) !important;
|
| 915 |
+
border: 2px solid var(--border-color) !important;
|
| 916 |
+
transition: border-color 0.3s ease !important;
|
| 917 |
+
}
|
| 918 |
+
|
| 919 |
+
.gr-textbox textarea:focus, .gr-textbox input:focus {
|
| 920 |
+
border-color: var(--primary-color) !important;
|
| 921 |
+
box-shadow: 0 0 0 3px rgba(79, 70, 229, 0.1) !important;
|
| 922 |
+
}
|
| 923 |
+
|
| 924 |
+
.example-box {
|
| 925 |
+
display: none; /* removed tip/example boxes */
|
| 926 |
+
}
|
| 927 |
+
"""
|
| 928 |
+
) as demo:
|
| 929 |
|
| 930 |
+
# --- Main Container ---
|
| 931 |
+
with gr.Column(elem_classes="main-content"):
|
| 932 |
+
|
| 933 |
+
# --- Header ---
|
| 934 |
+
gr.HTML("""
|
| 935 |
+
<div class="app-header">
|
| 936 |
+
<h1>🚀 CPU-Optimized Document QA System</h1>
|
| 937 |
+
<p>Clean, Concise Answers from Your Documents</p>
|
| 938 |
+
</div>
|
| 939 |
+
""")
|
| 940 |
+
|
| 941 |
+
# --- Main Content Area ---
|
| 942 |
with gr.Row():
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 943 |
|
| 944 |
+
# --- Left Column: Inputs ---
|
| 945 |
+
with gr.Column(scale=1):
|
| 946 |
+
with gr.Column(elem_classes="input-container"):
|
| 947 |
+
with gr.Tabs():
|
| 948 |
+
|
| 949 |
+
# --- Hackathon Submission Tab ---
|
| 950 |
+
with gr.Tab("🎯 Hackathon Submission", id=0):
|
| 951 |
+
with gr.Column(elem_classes="tab-content"):
|
| 952 |
+
gr.HTML('<h3 class="section-title">📄 Document Analysis Setup</h3>')
|
| 953 |
+
|
| 954 |
+
hack_url = gr.Textbox(
|
| 955 |
+
label="📄 Document URL (PDF/DOCX)",
|
| 956 |
+
placeholder="Enter the public URL of the document...",
|
| 957 |
+
lines=2,
|
| 958 |
+
info="Supports PDF and DOCX formats from public URLs"
|
| 959 |
+
)
|
| 960 |
+
|
| 961 |
+
hack_questions = gr.Textbox(
|
| 962 |
+
label="❓ Questions (JSON array or one per line)",
|
| 963 |
+
placeholder='["What is the grace period?", "Is maternity covered?"]',
|
| 964 |
+
lines=8,
|
| 965 |
+
info="Enter questions as JSON array or one question per line"
|
| 966 |
+
)
|
| 967 |
+
|
| 968 |
+
with gr.Row():
|
| 969 |
+
hack_clear_btn = gr.Button("🗑️ Clear", variant="secondary", size="sm")
|
| 970 |
+
hack_submit_btn = gr.Button("🚀 Process Submission", variant="primary", size="lg")
|
| 971 |
+
|
| 972 |
+
# --- Single Query Analysis Tab ---
|
| 973 |
+
with gr.Tab("🔍 Single Query Analysis", id=1):
|
| 974 |
+
with gr.Column(elem_classes="tab-content"):
|
| 975 |
+
gr.HTML('<h3 class="section-title">🔍 Detailed Document Query</h3>')
|
| 976 |
+
|
| 977 |
+
single_url = gr.Textbox(
|
| 978 |
+
label="📄 Document URL",
|
| 979 |
+
placeholder="Enter the public URL of the document...",
|
| 980 |
+
lines=2,
|
| 981 |
+
info="URL to your PDF or DOCX document"
|
| 982 |
+
)
|
| 983 |
+
|
| 984 |
+
single_question = gr.Textbox(
|
| 985 |
+
label="❓ Your Question",
|
| 986 |
+
placeholder="What is the waiting period for cataract surgery?",
|
| 987 |
+
lines=5,
|
| 988 |
+
info="Ask a specific question about your document"
|
| 989 |
+
)
|
| 990 |
+
|
| 991 |
+
with gr.Row():
|
| 992 |
+
single_clear_btn = gr.Button("🗑️ Clear", variant="secondary", size="sm")
|
| 993 |
+
single_submit_btn = gr.Button("🔍 Get Detailed Answer", variant="primary", size="lg")
|
| 994 |
+
|
| 995 |
+
# --- Right Column: Outputs ---
|
| 996 |
+
with gr.Column(scale=2):
|
| 997 |
+
with gr.Column(elem_classes="output-container"):
|
| 998 |
+
gr.HTML('<h3 class="section-title">📊 Analysis Results</h3>')
|
| 999 |
+
|
| 1000 |
+
with gr.Tabs():
|
| 1001 |
+
with gr.Tab("✅ Hackathon Results", id=2):
|
| 1002 |
+
hack_output = gr.Textbox(
|
| 1003 |
+
label="📊 Hackathon JSON Response",
|
| 1004 |
+
lines=25,
|
| 1005 |
+
max_lines=35,
|
| 1006 |
+
interactive=False,
|
| 1007 |
+
info="Clean JSON response with concise answers",
|
| 1008 |
+
show_copy_button=True
|
| 1009 |
+
)
|
| 1010 |
+
|
| 1011 |
+
with gr.Tab("🔍 Single Query Results", id=3):
|
| 1012 |
+
single_output = gr.Textbox(
|
| 1013 |
+
label="📋 Detailed Single Query Response",
|
| 1014 |
+
lines=25,
|
| 1015 |
+
max_lines=35,
|
| 1016 |
+
interactive=False,
|
| 1017 |
+
info="Comprehensive answer with supporting context",
|
| 1018 |
+
show_copy_button=True
|
| 1019 |
+
)
|
| 1020 |
+
|
| 1021 |
|
| 1022 |
+
# Hackathon Tab Logic
|
| 1023 |
+
hack_submit_btn.click(
|
| 1024 |
+
fn=hackathon_wrapper,
|
| 1025 |
inputs=[hack_url, hack_questions],
|
| 1026 |
outputs=[hack_output]
|
| 1027 |
)
|
| 1028 |
|
| 1029 |
+
hack_clear_btn.click(
|
| 1030 |
+
lambda: (None, None, None),
|
| 1031 |
+
outputs=[hack_url, hack_questions, hack_output]
|
| 1032 |
+
)
|
| 1033 |
+
|
| 1034 |
+
# Single Query Tab Logic
|
| 1035 |
+
single_submit_btn.click(
|
| 1036 |
+
fn=single_query_wrapper,
|
| 1037 |
inputs=[single_url, single_question],
|
| 1038 |
outputs=[single_output]
|
| 1039 |
)
|
| 1040 |
+
|
| 1041 |
+
single_clear_btn.click(
|
| 1042 |
+
lambda: (None, None, None),
|
| 1043 |
+
outputs=[single_url, single_question, single_output]
|
| 1044 |
+
)
|
| 1045 |
|
| 1046 |
# Queue for better performance on Spaces
|
| 1047 |
demo.queue(max_size=5)
|