File size: 22,586 Bytes
a213258
 
 
 
028ef27
a213258
 
 
 
 
 
 
 
 
 
 
 
028ef27
 
 
a213258
 
58c21ea
a213258
 
 
 
c4ddc69
a213258
58c21ea
a213258
 
 
 
 
 
58c21ea
 
a213258
c4ddc69
a213258
 
 
c4ddc69
 
 
58c21ea
 
 
 
 
 
 
c4ddc69
 
 
 
58c21ea
 
 
 
 
 
 
 
 
 
c4ddc69
58c21ea
 
c4ddc69
58c21ea
c4ddc69
58c21ea
c4ddc69
58c21ea
 
 
c4ddc69
58c21ea
c4ddc69
 
 
 
 
58c21ea
 
 
 
 
c4ddc69
58c21ea
c4ddc69
 
 
 
58c21ea
c4ddc69
58c21ea
 
a213258
c4ddc69
 
58c21ea
c4ddc69
 
 
 
a213258
c4ddc69
 
a213258
c4ddc69
 
 
a213258
c4ddc69
 
58c21ea
c4ddc69
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a213258
c4ddc69
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
58c21ea
c4ddc69
 
 
 
 
58c21ea
c4ddc69
58c21ea
028ef27
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a213258
 
 
 
028ef27
a213258
028ef27
 
 
 
 
 
 
 
a213258
 
028ef27
a213258
028ef27
 
 
a213258
 
 
 
028ef27
a213258
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
028ef27
a213258
 
 
 
 
028ef27
 
a213258
 
 
 
 
028ef27
a213258
 
 
 
 
 
 
 
 
 
 
 
 
028ef27
a213258
 
028ef27
a213258
 
 
 
 
 
 
 
c4ddc69
a213258
 
 
58c21ea
c4ddc69
58c21ea
a213258
58c21ea
 
a213258
58c21ea
 
 
 
a213258
c4ddc69
a213258
 
58c21ea
 
a213258
c4ddc69
a213258
 
c4ddc69
028ef27
a213258
028ef27
 
c4ddc69
 
a213258
028ef27
a213258
 
c4ddc69
52efc05
a213258
c4ddc69
 
a213258
c4ddc69
 
58c21ea
c4ddc69
 
a213258
028ef27
c4ddc69
 
 
 
 
 
a213258
 
 
028ef27
a213258
 
 
 
 
 
 
 
 
028ef27
a213258
 
 
 
 
028ef27
a213258
 
 
 
 
 
028ef27
 
 
a213258
 
028ef27
 
 
 
 
a213258
028ef27
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a213258
 
 
 
028ef27
a213258
028ef27
a213258
028ef27
a213258
 
 
028ef27
a213258
 
 
 
 
 
 
 
 
 
 
 
 
4a200c6
a213258
 
 
 
 
 
 
 
 
 
 
 
 
 
028ef27
a213258
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
aa67ef7
a213258
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
028ef27
a213258
 
 
 
 
 
 
 
 
 
 
 
028ef27
 
932c864
a213258
028ef27
a213258
 
 
 
 
 
028ef27
a213258
 
 
 
 
 
 
 
 
028ef27
a213258
 
 
 
 
028ef27
 
 
932c864
a213258
 
 
 
 
 
 
 
028ef27
a213258
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
import gradio as gr
import os
import json
import re
from typing import Iterator, Dict, Any, List, Optional
from openai import OpenAI
from openai.types.chat import ChatCompletionChunk

# Load abstracts content once at startup
def load_abstracts_content():
    """Load the abstracts content once at startup to avoid reading file on every request."""
    try:
        with open("abstracts.md", "r", encoding="utf-8") as f:
            return f.read()
    except FileNotFoundError:
        return "Abstracts database not found."

# Load abstracts content globally
ABSTRACTS_CONTENT = load_abstracts_content()

# Load full paper texts
def load_paper_texts():
    """Load all paper texts from the Papers directory and create a mapping from abstracts filenames."""
    papers = {}
    papers_dir = "Papers"
    
    if not os.path.exists(papers_dir):
        return {}
    
    # Create a mapping from abstracts filenames to actual file content
    for filename in os.listdir(papers_dir):
        if filename.endswith('.txt'):
            filepath = os.path.join(papers_dir, filename)
            try:
                with open(filepath, "r", encoding="utf-8") as f:
                    content = f.read()
                    # Store with the filename as key
                    papers[filename] = content
            except Exception as e:
                papers[filename] = f"Error loading paper: {str(e)}"
    
    return papers

# Load paper texts globally
PAPER_TEXTS = load_paper_texts()

def normalize_filename(filename):
    """Normalize filename for better matching."""
    # Remove .txt extension and normalize
    if filename.endswith('.txt'):
        filename = filename[:-4]
    
    # Convert to lowercase and remove special characters
    filename = re.sub(r'[^\w\s]', '', filename.lower())
    # Normalize whitespace
    filename = ' '.join(filename.split())
    return filename

def find_matching_paper_file(query_terms, papers_dict):
    """Find the best matching paper file based on query terms."""
    query_normalized = normalize_filename(' '.join(query_terms))
    best_match = None
    best_score = 0
    
    for filename in papers_dict.keys():
        filename_normalized = normalize_filename(filename)
        
        # Calculate match score
        score = 0
        
        # Exact substring match
        if query_normalized in filename_normalized or filename_normalized in query_normalized:
            score += 10
        
        # Word overlap
        query_words = set(query_normalized.split())
        filename_words = set(filename_normalized.split())
        overlap = len(query_words.intersection(filename_words))
        score += overlap * 2
        
        # Partial word matches
        for query_word in query_words:
            for filename_word in filename_words:
                if query_word in filename_word or filename_word in query_word:
                    score += 1
        
        if score > best_score:
            best_score = score
            best_match = filename
    
    return best_match if best_score > 0 else None

def get_relevant_papers_content(query, max_papers=5):
    """Get relevant paper content based on user query."""
    query_terms = query.lower().split()
    relevant_papers = []
    
    for filename, content in PAPER_TEXTS.items():
        title = filename[:-4] if filename.endswith('.txt') else filename
        title_lower = title.lower()
        
        # Calculate relevance score
        score = 0
        for term in query_terms:
            if term in title_lower:
                score += 2
            if term in content.lower():
                score += 1
        
        if score > 0:
            relevant_papers.append((filename, content, score))
    
    # Sort by relevance score and return top papers
    relevant_papers.sort(key=lambda x: x[2], reverse=True)
    return relevant_papers[:max_papers]

def get_full_paper_content(title, max_chars=12000):
    """Get full paper content for a specific title."""
    for filename, content in PAPER_TEXTS.items():
        if title.lower() in filename.lower() or filename.lower() in title.lower():
            return content[:max_chars] + "..." if len(content) > max_chars else content
    return "Paper not found."

def get_paper_summary(title):
    """Get a structured summary of a paper."""
    content = get_full_paper_content(title)
    if content == "Paper not found.":
        return content
    
    # Extract key sections
    sections = {
        'abstract': '',
        'introduction': '',
        'methodology': '',
        'results': '',
        'conclusions': ''
    }
    
    lines = content.split('\n')
    current_section = None
    
    for line in lines:
        line_lower = line.lower().strip()
        
        # Detect section headers
        if any(keyword in line_lower for keyword in ['abstract', 'introduction', 'method', 'methodology', 'results', 'conclusion']):
            if 'abstract' in line_lower:
                current_section = 'abstract'
            elif 'introduction' in line_lower:
                current_section = 'introduction'
            elif 'method' in line_lower:
                current_section = 'methodology'
            elif 'result' in line_lower:
                current_section = 'results'
            elif 'conclusion' in line_lower:
                current_section = 'conclusions'
        
        # Add content to current section
        if current_section and line.strip():
            sections[current_section] += line + '\n'
    
    # Create structured summary
    summary = f"# {title}\n\n"
    for section, content in sections.items():
        if content.strip():
            summary += f"## {section.title()}\n{content.strip()}\n\n"
    
    return summary

# Get API key with better error handling
api_key = os.getenv("OPENAI_API_KEY")
if not api_key:
    print("โš ๏ธ  Warning: OPENAI_API_KEY environment variable not set!")
    client = None
else:
    client = OpenAI(
        api_key=api_key,
        timeout=60.0,
        max_retries=3
    )

# Available models
AVAILABLE_MODELS = {
    "GPT-4o-mini": "gpt-4o-mini",
    "GPT-4o": "gpt-4o",
    "GPT-3.5 Turbo": "gpt-3.5-turbo"
}

# Define the tool for fetching papers
FETCH_PAPERS_TOOL = {
    "type": "function",
    "function": {
        "name": "fetch_papers",
        "description": "Fetch full text content of research papers by their filenames. Use this when you need detailed information, full text, conclusions, methodology, or specific quotes from papers.",
        "parameters": {
            "type": "object",
            "properties": {
                "filenames": {
                    "type": "array",
                    "items": {
                        "type": "string"
                    },
                    "description": "List of paper filenames to fetch (e.g., ['The Labor Market Effects of Generativ.txt', 'AI Companions Reduce Loneliness.txt'])"
                }
            },
            "required": ["filenames"]
        }
    }
}

def fetch_papers(filenames: List[str]) -> Dict[str, str]:
    """
    Fetch full paper texts by filenames.
    Returns a dictionary mapping filename to content.
    """
    papers = {}
    papers_dir = "Papers"
    
    if not os.path.exists(papers_dir):
        return {"error": "Papers directory not found"}
    
    for filename in filenames:
        # Ensure .txt extension
        if not filename.endswith('.txt'):
            filename += '.txt'
            
        filepath = os.path.join(papers_dir, filename)
        
        if os.path.exists(filepath):
            try:
                with open(filepath, "r", encoding="utf-8") as f:
                    papers[filename] = f.read()
            except Exception as e:
                papers[filename] = f"Error loading paper: {str(e)}"
        else:
            papers[filename] = f"Paper not found: {filename}"
    
    return papers

def extract_conclusion_from_paper(content: str) -> str:
    """Extract the conclusion section from a paper's content."""
    conclusion_patterns = [
        "conclusion and future works",
        "conclusion and future work", 
        "conclusions",
        "conclusion",
        "summary and conclusions",
        "discussion and conclusions"
    ]
    
    lines = content.split('\n')
    conclusion_start = -1
    
    for i, line in enumerate(lines):
        line_lower = line.lower().strip()
        if any(pattern in line_lower for pattern in conclusion_patterns):
            if (line.isupper() or 
                line.strip().endswith(':') or 
                len(line.strip()) < 100 or
                line.strip().startswith('Conclusion')):
                conclusion_start = i
                break
    
    if conclusion_start != -1:
        conclusion_lines = []
        for line in lines[conclusion_start:]:
            line_stripped = line.strip()
            if (line_stripped.lower().startswith('acknowledgments') or 
                line_stripped.lower().startswith('references') or
                line_stripped.startswith('--- Page')):
                break
            conclusion_lines.append(line)
        
        return '\n'.join(conclusion_lines)
    
    # Fallback: return the last 1000 characters
    return content[-1000:] if len(content) > 1000 else content

def truncate_conversation_history(messages: list, max_tokens: int = 8000) -> list:
    """Truncate conversation history to stay within token limits."""
    if len(messages) <= 3:
        return messages
    
    system_message = messages[0]
    conversation_messages = messages[1:]
    
    while len(conversation_messages) > 6:
        conversation_messages = conversation_messages[2:]
    
    return [system_message] + conversation_messages

def respond(
    message: str,
    history: list[tuple[str, str]],
    model_name: str,
    max_tokens: int,
    temperature: float,
    top_p: float,
) -> Iterator[str]:
    """
    Generate a response using OpenAI's models with function calling.
    """
    if not client:
        yield "โŒ Error: OpenAI API key not configured."
        return
    
    if not message.strip():
        yield "Please enter a message to start the conversation."
        return
    
    # Get relevant full paper content based on user query
    relevant_papers_content = get_relevant_papers_content(message)

    # Check if user is asking for a specific paper (e.g., "show me the full paper about pigs")
    specific_paper_content = ""
    conclusion_content = ""
    paper_summary_content = ""

    if any(keyword in message.lower() for keyword in ["full paper", "complete paper", "entire paper", "show me the paper", "read the paper", "summarize", "summary"]):
        # Try to find specific paper content
        for filename, content in PAPER_TEXTS.items():
            title = filename[:-4] if filename.endswith('.txt') else filename
            if any(term in title.lower() for term in message.lower().split()):
                if any(keyword in message.lower() for keyword in ["summarize", "summary"]):
                    paper_summary_content = get_paper_summary(title)
                else:
                    specific_paper_content = get_full_paper_content(title)
                break

    # Check if user is asking for conclusions specifically
    if any(keyword in message.lower() for keyword in ["conclusion", "conclusions", "what's the conclusion", "what is the conclusion"]):
        for filename, content in PAPER_TEXTS.items():
            title = filename[:-4] if filename.endswith('.txt') else filename
            if any(term in title.lower() for term in message.lower().split()):
                conclusion_content = extract_conclusion_from_paper(content)
                break

    # Initialize messages with a comprehensive system prompt
    system_prompt = f"""You are an AI chatbot designed to help users explore and analyze AI research papers. 

You have access to:
1. An abstracts database with summaries of research papers
2. Full paper texts for detailed analysis
3. A tool to fetch additional paper content when needed

ABSTRACTS DATABASE:
{ABSTRACTS_CONTENT}

RELEVANT PAPERS CONTENT:
{chr(10).join([f"Paper: {filename}{chr(10)}Content: {content[:3000]}..." for filename, content, score in relevant_papers_content])}

SPECIFIC PAPER CONTENT:
{specific_paper_content if specific_paper_content else "None"}

CONCLUSION CONTENT:
{conclusion_content if conclusion_content else "None"}

PAPER SUMMARY:
{paper_summary_content if paper_summary_content else "None"}

INSTRUCTIONS:
- Use the abstracts for general questions and overview
- Use full paper content when users ask for specific details, conclusions, or complete papers
- Use the fetch_papers tool when you need additional paper content
- Provide accurate, detailed responses based on the actual paper content
- When referencing papers, use their actual titles from the filenames
- Prioritize full paper content over abstracts when available"""
    
    messages = [{"role": "system", "content": system_prompt}]
    
    # Add conversation history
    for user_msg, assistant_msg in history:
        if user_msg and user_msg.strip():
            messages.append({"role": "user", "content": user_msg.strip()})
        if assistant_msg and assistant_msg.strip():
            messages.append({"role": "assistant", "content": assistant_msg.strip()})
    
    # Add current user message
    messages.append({"role": "user", "content": message.strip()})
    
    # Truncate if needed
    messages = truncate_conversation_history(messages)
    
    try:
        model = AVAILABLE_MODELS.get(model_name, "gpt-4o-mini")
        
        # Initial response with tool support
        response = client.chat.completions.create(
            model=model,
            messages=messages,
            max_tokens=max_tokens,
            temperature=temperature,
            top_p=top_p,
            tools=[FETCH_PAPERS_TOOL],
            tool_choice="auto",
            stream=True
        )
        
        # Collect the response and handle tool calls
        full_response = ""
        tool_calls = []
        current_tool_call = None
        
        for chunk in response:
            if hasattr(chunk.choices[0], 'delta'):
                delta = chunk.choices[0].delta
                
                # Handle regular content
                if delta.content is not None:
                    full_response += delta.content
                    yield full_response
                
                # Handle tool calls
                if delta.tool_calls:
                    for tool_call_chunk in delta.tool_calls:
                        if tool_call_chunk.id:
                            # New tool call
                            if current_tool_call:
                                tool_calls.append(current_tool_call)
                            current_tool_call = {
                                "id": tool_call_chunk.id,
                                "type": "function",
                                "function": {
                                    "name": tool_call_chunk.function.name if tool_call_chunk.function else "",
                                    "arguments": ""
                                }
                            }
                        
                        if current_tool_call and tool_call_chunk.function:
                            if tool_call_chunk.function.arguments:
                                current_tool_call["function"]["arguments"] += tool_call_chunk.function.arguments
        
        # Add final tool call if exists
        if current_tool_call:
            tool_calls.append(current_tool_call)
        
        # Process tool calls if any
        if tool_calls:
            # Add the assistant's message with tool calls
            messages.append({
                "role": "assistant",
                "content": full_response if full_response else None,
                "tool_calls": tool_calls
            })
            
            # Execute tool calls
            for tool_call in tool_calls:
                function_name = tool_call["function"]["name"]
                
                if function_name == "fetch_papers":
                    try:
                        # Parse arguments
                        arguments = json.loads(tool_call["function"]["arguments"])
                        filenames = arguments.get("filenames", [])
                        
                        # Fetch papers
                        papers_content = fetch_papers(filenames)
                        
                        # Add tool response to messages
                        tool_response = {
                            "role": "tool",
                            "tool_call_id": tool_call["id"],
                            "content": json.dumps(papers_content)
                        }
                        messages.append(tool_response)
                        
                    except Exception as e:
                        tool_response = {
                            "role": "tool",
                            "tool_call_id": tool_call["id"],
                            "content": f"Error: {str(e)}"
                        }
                        messages.append(tool_response)
            
            # Get final response with tool results
            final_response = client.chat.completions.create(
                model=model,
                messages=messages,
                max_tokens=max_tokens,
                temperature=temperature,
                top_p=top_p,
                stream=True
            )
            
            # Stream the final response
            final_text = ""
            for chunk in final_response:
                if hasattr(chunk.choices[0], 'delta') and chunk.choices[0].delta.content is not None:
                    final_text += chunk.choices[0].delta.content
                    yield full_response + "\n\n" + final_text if full_response else final_text
                
    except Exception as e:
        error_message = f"Error: {str(e)}"
        if "api_key" in str(e).lower():
            error_message = "Error: Invalid or missing OpenAI API key."
        elif "quota" in str(e).lower():
            error_message = "Error: API quota exceeded."
        elif "rate" in str(e).lower():
            error_message = "Error: Rate limit exceeded."
        yield error_message

def chat_fn(message, history, model_name, max_tokens, temperature, top_p):
    """Handle the entire chat interaction."""
    if not message.strip():
        return history
    
    history.append([message, ""])
    
    for response in respond(message, history[:-1], model_name, max_tokens, temperature, top_p):
        history[-1][1] = response
        yield history

def clear_history() -> tuple:
    """Clear the conversation history."""
    return [], ""

        # Create the Gradio interface
with gr.Blocks(
    title="๐Ÿ“š AI Research Paper Chatbot",
    theme=gr.themes.Soft(),
    css="""
    .gradio-container {
        max-width: 1200px !important;
        margin: auto !important;
    }
    """
) as demo:
    gr.Markdown(
        """
        # ๐Ÿ“š AI Research Paper Chatbot
        
        Chat with an AI assistant that can intelligently retrieve and analyze research papers.
        """
    )
    
    with gr.Row():
        with gr.Column(scale=3):
            chatbot = gr.Chatbot(
                height=500,
                show_label=False,
                container=True,
                bubble_full_width=False
            )
            
            with gr.Row():
                msg = gr.Textbox(
                    placeholder="Type your message here...",
                    show_label=False,
                    container=False,
                    scale=9
                )
                submit_btn = gr.Button("Send", variant="primary", scale=1)
                clear_btn = gr.Button("Clear", variant="secondary", scale=1)
        
        with gr.Column(scale=1):
            gr.Markdown("### โš™๏ธ Settings")
            
            model_dropdown = gr.Dropdown(
                choices=list(AVAILABLE_MODELS.keys()),
                value="GPT-4o",
                label="Model",
                info="Select the AI model to use"
            )
            
            max_tokens_slider = gr.Slider(
                minimum=1,
                maximum=4096,
                value=1024,
                step=1,
                label="Max Tokens",
                info="Maximum response length"
            )
            
            temperature_slider = gr.Slider(
                minimum=0.0,
                maximum=2.0,
                value=0.7,
                step=0.1,
                label="Temperature",
                info="Creativity level"
            )
            
            top_p_slider = gr.Slider(
                minimum=0.0,
                maximum=1.0,
                value=1.0,
                step=0.05,
                label="Top-p",
                info="Response diversity"
            )
            
            gr.Markdown("### ๐Ÿ’ก Examples")
            example_btn1 = gr.Button("What papers discuss AI's impact on employment?", size="sm")
            example_btn2 = gr.Button("Show me the full paper about AI companions", size="sm")
            example_btn3 = gr.Button("Compare findings on AI in education", size="sm")
    
    # Event handlers
    msg.submit(
        chat_fn,
        [msg, chatbot, model_dropdown, max_tokens_slider, temperature_slider, top_p_slider],
        [chatbot],
        show_progress=True
    ).then(
        lambda: "",
        outputs=[msg]
    )
    
    submit_btn.click(
        chat_fn,
        [msg, chatbot, model_dropdown, max_tokens_slider, temperature_slider, top_p_slider],
        [chatbot],
        show_progress=True
    ).then(
        lambda: "",
        outputs=[msg]
    )
    
    clear_btn.click(clear_history, outputs=[chatbot, msg])
    
    # Example handlers
    example_btn1.click(lambda: "What papers discuss AI's impact on employment?", outputs=msg)
    example_btn2.click(lambda: "Show me the full paper about AI companions", outputs=msg)
    example_btn3.click(lambda: "Compare findings on AI in education", outputs=msg)

if __name__ == "__main__":
    if not os.getenv("OPENAI_API_KEY"):
        print("โš ๏ธ  Warning: OPENAI_API_KEY environment variable not set!")
    
    demo.launch(
        server_name="0.0.0.0",
        server_port=7860,
        share=False,
        show_error=True,
        quiet=False
    )