Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -48,12 +48,7 @@ quant_config = BitsAndBytesConfig(
|
|
| 48 |
HF_TOKEN = os.getenv("HF_TOKEN")
|
| 49 |
|
| 50 |
def load_models():
|
| 51 |
-
ner_tok = AutoTokenizer.from_pretrained(
|
| 52 |
-
NER_ID,
|
| 53 |
-
token=HF_TOKEN,
|
| 54 |
-
use_fast=False
|
| 55 |
-
)
|
| 56 |
-
|
| 57 |
ner_mod = AutoModelForTokenClassification.from_pretrained(NER_ID, token=HF_TOKEN)
|
| 58 |
ner_mod.eval()
|
| 59 |
if torch.cuda.is_available():
|
|
@@ -61,9 +56,8 @@ def load_models():
|
|
| 61 |
|
| 62 |
base_mod = AutoModelForCausalLM.from_pretrained(
|
| 63 |
BASE_ID,
|
| 64 |
-
|
| 65 |
-
|
| 66 |
-
offload_folder="/tmp/offload"
|
| 67 |
)
|
| 68 |
norm_tok = AutoTokenizer.from_pretrained(ADAPTER_ID, use_fast=True, token=HF_TOKEN)
|
| 69 |
norm_mod = PeftModel.from_pretrained(
|
|
@@ -234,5 +228,4 @@ with gr.Blocks() as demo:
|
|
| 234 |
)
|
| 235 |
|
| 236 |
# Lanzar la app
|
| 237 |
-
demo.launch()
|
| 238 |
-
|
|
|
|
| 48 |
HF_TOKEN = os.getenv("HF_TOKEN")
|
| 49 |
|
| 50 |
def load_models():
|
| 51 |
+
ner_tok = AutoTokenizer.from_pretrained(NER_ID, token=HF_TOKEN)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 52 |
ner_mod = AutoModelForTokenClassification.from_pretrained(NER_ID, token=HF_TOKEN)
|
| 53 |
ner_mod.eval()
|
| 54 |
if torch.cuda.is_available():
|
|
|
|
| 56 |
|
| 57 |
base_mod = AutoModelForCausalLM.from_pretrained(
|
| 58 |
BASE_ID,
|
| 59 |
+
device_map="auto",
|
| 60 |
+
token=HF_TOKEN
|
|
|
|
| 61 |
)
|
| 62 |
norm_tok = AutoTokenizer.from_pretrained(ADAPTER_ID, use_fast=True, token=HF_TOKEN)
|
| 63 |
norm_mod = PeftModel.from_pretrained(
|
|
|
|
| 228 |
)
|
| 229 |
|
| 230 |
# Lanzar la app
|
| 231 |
+
demo.launch()
|
|
|