Upload app.py
Browse files
app.py
CHANGED
|
@@ -3,7 +3,6 @@ os.environ["TF_ENABLE_ONEDNN_OPTS"] = "0"
|
|
| 3 |
import streamlit as st
|
| 4 |
from pathlib import Path
|
| 5 |
import torch, json, csv, warnings
|
| 6 |
-
from transformers import T5Tokenizer, T5ForConditionalGeneration
|
| 7 |
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
| 8 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 9 |
from pathlib import Path
|
|
@@ -48,9 +47,6 @@ def guardar_interaccion_dual(pregunta, respuesta, tipo, user_id):
|
|
| 48 |
}
|
| 49 |
f_jsonl.write(json.dumps(registro, ensure_ascii=False) + "\n")
|
| 50 |
|
| 51 |
-
# ✅ Corrige la ruta correctamente desde Scripts hacia Models
|
| 52 |
-
def get_model_path(folder_name):
|
| 53 |
-
return Path("Models") / folder_name
|
| 54 |
|
| 55 |
# Función para cargar modelos
|
| 56 |
@st.cache_resource
|
|
|
|
| 3 |
import streamlit as st
|
| 4 |
from pathlib import Path
|
| 5 |
import torch, json, csv, warnings
|
|
|
|
| 6 |
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
| 7 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 8 |
from pathlib import Path
|
|
|
|
| 47 |
}
|
| 48 |
f_jsonl.write(json.dumps(registro, ensure_ascii=False) + "\n")
|
| 49 |
|
|
|
|
|
|
|
|
|
|
| 50 |
|
| 51 |
# Función para cargar modelos
|
| 52 |
@st.cache_resource
|