Update app.py
Browse files
app.py
CHANGED
|
@@ -14,7 +14,7 @@ model, tokenizer = FastLanguageModel.from_pretrained(
|
|
| 14 |
load_in_4bit = True,
|
| 15 |
token= READ_HF
|
| 16 |
)
|
| 17 |
-
|
| 18 |
|
| 19 |
|
| 20 |
alpaca_prompt = """Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
|
|
@@ -77,7 +77,7 @@ Category List : ["Dairy & Eggs", "Beverages & Snacks", "Cleaning & Hygiene", "Gr
|
|
| 77 |
|
| 78 |
@spaces.GPU()
|
| 79 |
def chunk_it(inventory_list, user_input_text):
|
| 80 |
-
|
| 81 |
inputs = tokenizer(
|
| 82 |
[
|
| 83 |
alpaca_prompt.format(
|
|
@@ -104,7 +104,7 @@ iface = gr.Interface(
|
|
| 104 |
gr.Textbox(label="user_input_text", lines=3),
|
| 105 |
gr.Textbox(label="inventory_list", lines=5)
|
| 106 |
],
|
| 107 |
-
outputs="
|
| 108 |
title="Testing",
|
| 109 |
)
|
| 110 |
|
|
|
|
| 14 |
load_in_4bit = True,
|
| 15 |
token= READ_HF
|
| 16 |
)
|
| 17 |
+
|
| 18 |
|
| 19 |
|
| 20 |
alpaca_prompt = """Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
|
|
|
|
| 77 |
|
| 78 |
@spaces.GPU()
|
| 79 |
def chunk_it(inventory_list, user_input_text):
|
| 80 |
+
FastLanguageModel.for_inference(model) # Enable native 2x faster inference
|
| 81 |
inputs = tokenizer(
|
| 82 |
[
|
| 83 |
alpaca_prompt.format(
|
|
|
|
| 104 |
gr.Textbox(label="user_input_text", lines=3),
|
| 105 |
gr.Textbox(label="inventory_list", lines=5)
|
| 106 |
],
|
| 107 |
+
outputs=gr.Textbox(label="output", lines=23),
|
| 108 |
title="Testing",
|
| 109 |
)
|
| 110 |
|