Update app.py
Browse files
app.py
CHANGED
|
@@ -6,6 +6,12 @@ from fpdf import FPDF
|
|
| 6 |
import tempfile
|
| 7 |
from io import BytesIO
|
| 8 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
|
| 10 |
# ------------------ MODEL LOADING ------------------
|
| 11 |
MODEL_NAME = "Qwen/Qwen2.5-1.5B-Instruct"
|
|
@@ -15,7 +21,7 @@ def load_llm():
|
|
| 15 |
model = AutoModelForCausalLM.from_pretrained(
|
| 16 |
MODEL_NAME,
|
| 17 |
device_map="auto",
|
| 18 |
-
torch_dtype="auto"
|
| 19 |
)
|
| 20 |
return tokenizer, model
|
| 21 |
|
|
@@ -139,7 +145,6 @@ def build_pdf(risk_icon, risk_label, summary, anomalies):
|
|
| 139 |
|
| 140 |
|
| 141 |
# ------------------ STREAMLIT UI ------------------
|
| 142 |
-
st.set_page_config(page_title="Smart Log Copilot", layout="wide")
|
| 143 |
st.title("🔍 Smart Log Copilot (CSV-powered LLM Demo)")
|
| 144 |
|
| 145 |
uploaded = st.file_uploader("Upload CSV log file", type=["csv"])
|
|
|
|
| 6 |
import tempfile
|
| 7 |
from io import BytesIO
|
| 8 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 9 |
+
import torch
|
| 10 |
+
|
| 11 |
+
# ---------------------------------------------------------------------
|
| 12 |
+
# MUST BE FIRST STREAMLIT CALL
|
| 13 |
+
# ---------------------------------------------------------------------
|
| 14 |
+
st.set_page_config(page_title="Smart Log Copilot", layout="wide")
|
| 15 |
|
| 16 |
# ------------------ MODEL LOADING ------------------
|
| 17 |
MODEL_NAME = "Qwen/Qwen2.5-1.5B-Instruct"
|
|
|
|
| 21 |
model = AutoModelForCausalLM.from_pretrained(
|
| 22 |
MODEL_NAME,
|
| 23 |
device_map="auto",
|
| 24 |
+
torch_dtype=torch.float16 if torch.cuda.is_available() else "auto"
|
| 25 |
)
|
| 26 |
return tokenizer, model
|
| 27 |
|
|
|
|
| 145 |
|
| 146 |
|
| 147 |
# ------------------ STREAMLIT UI ------------------
|
|
|
|
| 148 |
st.title("🔍 Smart Log Copilot (CSV-powered LLM Demo)")
|
| 149 |
|
| 150 |
uploaded = st.file_uploader("Upload CSV log file", type=["csv"])
|