Spaces:
Runtime error
Runtime error
Commit
·
bfc086f
1
Parent(s):
323ed1c
added anther model to test
Browse files
app.py
CHANGED
|
@@ -54,7 +54,7 @@ model_mapping = {
|
|
| 54 |
#'vicuna-7b': 'lmsys/vicuna-7b-v1.3',
|
| 55 |
#'vicuna-13b': 'lmsys/vicuna-13b-v1.3',
|
| 56 |
#'vicuna-33b': 'lmsys/vicuna-33b-v1.3',
|
| 57 |
-
|
| 58 |
#'llama-7b': './llama/hf/7B',
|
| 59 |
#'llama-13b': './llama/hf/13B',
|
| 60 |
#'llama-30b': './llama/hf/30B',
|
|
@@ -98,7 +98,7 @@ gpt_pipeline = pipeline(task="text-generation", model="gpt2")
|
|
| 98 |
#vicuna7b_pipeline = pipeline(task="text2text-generation", model="lmsys/vicuna-7b-v1.3")
|
| 99 |
#vicuna13b_pipeline = pipeline(task="text2text-generation", model="lmsys/vicuna-13b-v1.3")
|
| 100 |
#vicuna33b_pipeline = pipeline(task="text2text-generation", model="lmsys/vicuna-33b-v1.3")
|
| 101 |
-
|
| 102 |
#llama7b_pipeline = pipeline(task="text2text-generation", model="./llama/hf/7B")
|
| 103 |
#llama13b_pipeline = pipeline(task="text2text-generation", model="./llama/hf/13B")
|
| 104 |
#llama30b_pipeline = pipeline(task="text2text-generation", model="./llama/hf/30B")
|
|
@@ -145,6 +145,34 @@ def process_text(model_name, task, text):
|
|
| 145 |
response2 = gpt_pipeline(strategy2)[0]['generated_text']
|
| 146 |
response3 = gpt_pipeline(strategy3)[0]['generated_text']
|
| 147 |
return (response1, response2, response3)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 148 |
|
| 149 |
# Define prompts for each strategy based on the task
|
| 150 |
#strategy_prompts = {
|
|
|
|
| 54 |
#'vicuna-7b': 'lmsys/vicuna-7b-v1.3',
|
| 55 |
#'vicuna-13b': 'lmsys/vicuna-13b-v1.3',
|
| 56 |
#'vicuna-33b': 'lmsys/vicuna-33b-v1.3',
|
| 57 |
+
'fastchat-t5': 'lmsys/fastchat-t5-3b-v1.0',
|
| 58 |
#'llama-7b': './llama/hf/7B',
|
| 59 |
#'llama-13b': './llama/hf/13B',
|
| 60 |
#'llama-30b': './llama/hf/30B',
|
|
|
|
| 98 |
#vicuna7b_pipeline = pipeline(task="text2text-generation", model="lmsys/vicuna-7b-v1.3")
|
| 99 |
#vicuna13b_pipeline = pipeline(task="text2text-generation", model="lmsys/vicuna-13b-v1.3")
|
| 100 |
#vicuna33b_pipeline = pipeline(task="text2text-generation", model="lmsys/vicuna-33b-v1.3")
|
| 101 |
+
fastchatT5_pipeline = pipeline(task="text2text-generation", model="lmsys/fastchat-t5-3b-v1.0")
|
| 102 |
#llama7b_pipeline = pipeline(task="text2text-generation", model="./llama/hf/7B")
|
| 103 |
#llama13b_pipeline = pipeline(task="text2text-generation", model="./llama/hf/13B")
|
| 104 |
#llama30b_pipeline = pipeline(task="text2text-generation", model="./llama/hf/30B")
|
|
|
|
| 145 |
response2 = gpt_pipeline(strategy2)[0]['generated_text']
|
| 146 |
response3 = gpt_pipeline(strategy3)[0]['generated_text']
|
| 147 |
return (response1, response2, response3)
|
| 148 |
+
elif model_name == 'fastchat-t5':
|
| 149 |
+
if task == 'POS':
|
| 150 |
+
strategy1 = template_all.format(text)
|
| 151 |
+
strategy2 = prompt2_pos.format(text)
|
| 152 |
+
strategy3 = demon_pos
|
| 153 |
+
|
| 154 |
+
response1 = fastchatT5_pipeline(strategy1)[0]['generated_text']
|
| 155 |
+
response2 = fastchatT5_pipeline(strategy2)[0]['generated_text']
|
| 156 |
+
response3 = fastchatT5_pipeline(strategy3)[0]['generated_text']
|
| 157 |
+
return (response1, response2, response3)
|
| 158 |
+
elif task == 'Chunking':
|
| 159 |
+
strategy1 = template_all.format(text)
|
| 160 |
+
strategy2 = prompt2_chunk.format(text)
|
| 161 |
+
strategy3 = demon_chunk
|
| 162 |
+
|
| 163 |
+
response1 = fastchatT5_pipeline(strategy1)[0]['generated_text']
|
| 164 |
+
response2 = fastchatT5_pipeline(strategy2)[0]['generated_text']
|
| 165 |
+
response3 = fastchatT5_pipeline(strategy3)[0]['generated_text']
|
| 166 |
+
return (response1, response2, response3)
|
| 167 |
+
elif task == 'Parsing':
|
| 168 |
+
strategy1 = template_all.format(text)
|
| 169 |
+
strategy2 = prompt2_parse.format(text)
|
| 170 |
+
strategy3 = demon_parse
|
| 171 |
+
|
| 172 |
+
response1 = fastchatT5_pipeline(strategy1)[0]['generated_text']
|
| 173 |
+
response2 = fastchatT5_pipeline(strategy2)[0]['generated_text']
|
| 174 |
+
response3 = fastchatT5_pipeline(strategy3)[0]['generated_text']
|
| 175 |
+
return (response1, response2, response3)
|
| 176 |
|
| 177 |
# Define prompts for each strategy based on the task
|
| 178 |
#strategy_prompts = {
|