Spaces:
Sleeping
Sleeping
| import gradio as gr | |
| import os | |
| import time | |
| import GPTSimple as ai | |
| import random | |
| import vision | |
| # Chatbot demo with multimodal input (text, markdown, LaTeX, code blocks, image, audio, & video). Plus shows support for streaming text. | |
| DESC = "# LLaMA 3.1 Vision\n<p>LLaMA 3.1 Vision uses LLaMA 3.1 405B and Florence 2 to give vision to LLaMA</p>" | |
| def print_like_dislike(x: gr.LikeData): | |
| print(x.index, x.value, x.liked) | |
| def add_message(history, message): | |
| for x in message["files"]: | |
| history.append(((x,), None)) | |
| if message["text"] is not None: | |
| history.append((message["text"], None)) | |
| return history, gr.MultimodalTextbox(value=None, interactive=False) | |
| def bot(history): | |
| his = [{"role": "system", "content": "you are a helpful assistant. you can\"see\" image that the user sends by the description being in [IMG][/IMG]. don't reference how you can only see a description"}] | |
| nextone = "" | |
| for i in history: | |
| if isinstance(i[0], tuple): | |
| nextone += "[IMG]" + vision.see_file(i[0][0]) + "[/IMG]\n" | |
| else: | |
| his.append({"role": "user", "content": nextone + i[0]}) | |
| nextone = "" | |
| if i[1] is not None: | |
| his.append({"role": "assistant", "content": i[1]}) | |
| chat = ai.conversation(base_url="deepinfra", model="meta-llama/Meta-Llama-3.1-405B-Instruct", history=his) | |
| print(his) | |
| stre = chat.generate(stream=True) | |
| history[-1][1] = "" | |
| for character in stre: | |
| if character.token is not None: | |
| history[-1][1] += character.token | |
| yield history | |
| def clear_history(): | |
| return [], {"text":"", "files":[]} | |
| def retry_last(history): | |
| history[-1][1]=None | |
| res = bot(history) | |
| for i in res: | |
| yield i | |
| with gr.Blocks(fill_height=True, theme=gr.themes.Soft(), css="style.css") as demo: | |
| gr.Markdown(DESC) | |
| chatbot = gr.Chatbot( | |
| elem_id="chatbot", | |
| bubble_full_width=False, | |
| scale=1, | |
| show_label=False | |
| ) | |
| with gr.Row(): | |
| dl = gr.Button("", icon="delete.svg") | |
| chat_input = gr.MultimodalTextbox(interactive=True, | |
| file_count="multiple", | |
| placeholder="Enter message or upload file...", show_label=False) | |
| re = gr.Button("", icon="retry.svg") | |
| chat_msg = chat_input.submit(add_message, [chatbot, chat_input], [chatbot, chat_input]) | |
| bot_msg = chat_msg.then(bot, chatbot, chatbot, api_name="bot_response") | |
| bot_msg.then(lambda: gr.MultimodalTextbox(interactive=True), None, [chat_input]) | |
| dl.click(clear_history, None, [chatbot, chat_input]) | |
| re.click(retry_last, [chatbot], chatbot) | |
| demo.queue() | |
| demo.launch() | |