Update app.py
Browse files
app.py
CHANGED
|
@@ -68,7 +68,7 @@ Answer:"""
|
|
| 68 |
# Azure LLM setup
|
| 69 |
endpoint = "https://models.github.ai/inference"
|
| 70 |
model = "deepseek/DeepSeek-V3-0324"
|
| 71 |
-
token = os.
|
| 72 |
client = ChatCompletionsClient(endpoint=endpoint, credential=AzureKeyCredential(token))
|
| 73 |
|
| 74 |
# Main pipeline for Gradio
|
|
@@ -106,7 +106,7 @@ def handle_query(text_input, image_input):
|
|
| 106 |
# Gradio UI
|
| 107 |
def launch_app():
|
| 108 |
with gr.Blocks() as demo:
|
| 109 |
-
gr.Markdown("##
|
| 110 |
with gr.Row():
|
| 111 |
text_input = gr.Textbox(label="Enter your query")
|
| 112 |
image_input = gr.Image(label="Upload an image", type="pil")
|
|
|
|
| 68 |
# Azure LLM setup
|
| 69 |
endpoint = "https://models.github.ai/inference"
|
| 70 |
model = "deepseek/DeepSeek-V3-0324"
|
| 71 |
+
token = os.getenv("GITHUB_TOKEN")
|
| 72 |
client = ChatCompletionsClient(endpoint=endpoint, credential=AzureKeyCredential(token))
|
| 73 |
|
| 74 |
# Main pipeline for Gradio
|
|
|
|
| 106 |
# Gradio UI
|
| 107 |
def launch_app():
|
| 108 |
with gr.Blocks() as demo:
|
| 109 |
+
gr.Markdown("## 📄🎓Multimodal Chatbot for FAST-NUCES")
|
| 110 |
with gr.Row():
|
| 111 |
text_input = gr.Textbox(label="Enter your query")
|
| 112 |
image_input = gr.Image(label="Upload an image", type="pil")
|