Spaces:
Runtime error
Runtime error
Commit ·
49c7ae8
1
Parent(s): 35e0ec8
test
Browse files
app.py
CHANGED
|
@@ -7,6 +7,18 @@ model_name = "lmsys/vicuna-7b-v1.3"
|
|
| 7 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 8 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
| 9 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 10 |
with gr.Blocks() as demo:
|
| 11 |
gr.Markdown("# LLM Evaluator With Linguistic Scrutiny")
|
| 12 |
|
|
@@ -15,13 +27,13 @@ with gr.Blocks() as demo:
|
|
| 15 |
|
| 16 |
prompt_POS = gr.Textbox(show_label=False, placeholder="Write a prompt and press enter")
|
| 17 |
|
| 18 |
-
gr.Markdown("Strategy 1 QA")
|
| 19 |
with gr.Row():
|
| 20 |
vicuna_S1_chatbot_POS = gr.Chatbot(label="vicuna-7b")
|
| 21 |
llama_S1_chatbot_POS = gr.Chatbot(label="llama-7b")
|
| 22 |
gpt_S1_chatbot_POS = gr.Chatbot(label="gpt-3.5")
|
| 23 |
clear = gr.ClearButton([prompt_POS, vicuna_S1_chatbot_POS])
|
| 24 |
-
gr.Markdown("Strategy 2 Instruction")
|
| 25 |
with gr.Row():
|
| 26 |
vicuna_S2_chatbot_POS = gr.Chatbot(label="vicuna-7b")
|
| 27 |
llama_S2_chatbot_POS = gr.Chatbot(label="llama-7b")
|
|
@@ -67,12 +79,12 @@ with gr.Blocks() as demo:
|
|
| 67 |
time.sleep(2)
|
| 68 |
return "", chat_history
|
| 69 |
|
| 70 |
-
prompt_POS.submit(respond, [prompt_POS, vicuna_S1_chatbot_POS], [prompt_POS, vicuna_S1_chatbot_POS])
|
| 71 |
-
prompt_POS.submit(respond, [prompt_POS, vicuna_S2_chatbot_POS], [prompt_POS, vicuna_S2_chatbot_POS])
|
| 72 |
prompt_POS.submit(respond, [prompt_POS, vicuna_S3_chatbot_POS], [prompt_POS, vicuna_S3_chatbot_POS])
|
| 73 |
|
| 74 |
-
prompt_CHUNK.submit(respond, [prompt_CHUNK, vicuna_S1_chatbot_CHUNK], [prompt_CHUNK, vicuna_S1_chatbot_CHUNK])
|
| 75 |
-
prompt_CHUNK.submit(respond, [prompt_CHUNK, vicuna_S2_chatbot_CHUNK], [prompt_CHUNK, vicuna_S2_chatbot_CHUNK])
|
| 76 |
prompt_CHUNK.submit(respond, [prompt_CHUNK, vicuna_S3_chatbot_CHUNK], [prompt_CHUNK, vicuna_S3_chatbot_CHUNK])
|
| 77 |
|
| 78 |
demo.launch()
|
|
|
|
| 7 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 8 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
| 9 |
|
| 10 |
+
## Task 1
|
| 11 |
+
# msg = template_all.format(text)
|
| 12 |
+
template_all = '''Output the <Noun, Verb, Adjective, Adverb, Preposition/Subord, Coordinating Conjunction, Cardinal Number, Determiner, Noun Phrase, Verb Phrase, Adjective Phrase, Adverb Phrase, Preposition Phrase, Conjunction Phrase, Coordinate Phrase, Quantitave Phrase, Complex Nominal, Clause, Dependent Clause, Fragment Clause, T-unit, Complex T-unit, Fragment T-unit> in the following sentence without additional text in json format: "{}"'''
|
| 13 |
+
# msg = template_single.format(ents_prompt[eid], text)
|
| 14 |
+
template_single = '''Output any <{}> in the following sentence one per line without additional text: "{}"'''
|
| 15 |
+
|
| 16 |
+
## Task 2
|
| 17 |
+
prompt2_pos = '''POS tag the following sentence using Universal POS tag set without generating additional text: {}'''
|
| 18 |
+
prompt2_chunk = '''Do sentence chunking for the following sentence as in CoNLL 2000 shared task without generating addtional text: {}'''
|
| 19 |
+
|
| 20 |
+
## Task 3
|
| 21 |
+
|
| 22 |
with gr.Blocks() as demo:
|
| 23 |
gr.Markdown("# LLM Evaluator With Linguistic Scrutiny")
|
| 24 |
|
|
|
|
| 27 |
|
| 28 |
prompt_POS = gr.Textbox(show_label=False, placeholder="Write a prompt and press enter")
|
| 29 |
|
| 30 |
+
gr.Markdown("Strategy 1 QA-Based Prompting")
|
| 31 |
with gr.Row():
|
| 32 |
vicuna_S1_chatbot_POS = gr.Chatbot(label="vicuna-7b")
|
| 33 |
llama_S1_chatbot_POS = gr.Chatbot(label="llama-7b")
|
| 34 |
gpt_S1_chatbot_POS = gr.Chatbot(label="gpt-3.5")
|
| 35 |
clear = gr.ClearButton([prompt_POS, vicuna_S1_chatbot_POS])
|
| 36 |
+
gr.Markdown("Strategy 2 Instruction-Based Prompting")
|
| 37 |
with gr.Row():
|
| 38 |
vicuna_S2_chatbot_POS = gr.Chatbot(label="vicuna-7b")
|
| 39 |
llama_S2_chatbot_POS = gr.Chatbot(label="llama-7b")
|
|
|
|
| 79 |
time.sleep(2)
|
| 80 |
return "", chat_history
|
| 81 |
|
| 82 |
+
prompt_POS.submit(respond, [template_all.format(prompt_POS), vicuna_S1_chatbot_POS], [template_all.format(prompt_POS), vicuna_S1_chatbot_POS])
|
| 83 |
+
prompt_POS.submit(respond, [prompt2_pos.format(prompt_POS), vicuna_S2_chatbot_POS], [prompt2_pos.format(prompt_POS), vicuna_S2_chatbot_POS])
|
| 84 |
prompt_POS.submit(respond, [prompt_POS, vicuna_S3_chatbot_POS], [prompt_POS, vicuna_S3_chatbot_POS])
|
| 85 |
|
| 86 |
+
prompt_CHUNK.submit(respond, [template_all.format(prompt_CHUNK), vicuna_S1_chatbot_CHUNK], [template_all.format(prompt_CHUNK), vicuna_S1_chatbot_CHUNK])
|
| 87 |
+
prompt_CHUNK.submit(respond, [prompt2_chunk.format(prompt_CHUNK), vicuna_S2_chatbot_CHUNK], [prompt2_chunk.format(prompt_CHUNK), vicuna_S2_chatbot_CHUNK])
|
| 88 |
prompt_CHUNK.submit(respond, [prompt_CHUNK, vicuna_S3_chatbot_CHUNK], [prompt_CHUNK, vicuna_S3_chatbot_CHUNK])
|
| 89 |
|
| 90 |
demo.launch()
|