Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -5,6 +5,7 @@ import numpy as np
|
|
| 5 |
import json
|
| 6 |
import os
|
| 7 |
import uuid
|
|
|
|
| 8 |
import re
|
| 9 |
from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
|
| 10 |
from peft import PeftModel, AutoPeftModelForCausalLM
|
|
@@ -29,6 +30,7 @@ model.eval()
|
|
| 29 |
tokenizer = AutoTokenizer.from_pretrained(lora_path)
|
| 30 |
|
| 31 |
# === Updated Generate Answer Function ===
|
|
|
|
| 32 |
def generate_answer(json_data: Dict[str, Any], question: str) -> str:
|
| 33 |
"""
|
| 34 |
Generate answer using the fine-tuned model.
|
|
|
|
| 5 |
import json
|
| 6 |
import os
|
| 7 |
import uuid
|
| 8 |
+
import spaces
|
| 9 |
import re
|
| 10 |
from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
|
| 11 |
from peft import PeftModel, AutoPeftModelForCausalLM
|
|
|
|
| 30 |
tokenizer = AutoTokenizer.from_pretrained(lora_path)
|
| 31 |
|
| 32 |
# === Updated Generate Answer Function ===
|
| 33 |
+
@spaces.GPU(duration=get_duration)
|
| 34 |
def generate_answer(json_data: Dict[str, Any], question: str) -> str:
|
| 35 |
"""
|
| 36 |
Generate answer using the fine-tuned model.
|