Update app.py
Browse files
app.py
CHANGED
|
@@ -170,7 +170,7 @@ TOOLS = [
|
|
| 170 |
),
|
| 171 |
]
|
| 172 |
|
| 173 |
-
MODEL_NAME = "
|
| 174 |
model = None
|
| 175 |
tokenizer = None
|
| 176 |
model_loaded = False
|
|
@@ -185,6 +185,9 @@ def download_and_load_model(progress=gr.Progress()):
|
|
| 185 |
progress(0.2, desc="Downloading tokenizer...")
|
| 186 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
|
| 187 |
|
|
|
|
|
|
|
|
|
|
| 188 |
progress(0.4, desc="Downloading model (this may take several minutes)...")
|
| 189 |
model = AutoModelForCausalLM.from_pretrained(
|
| 190 |
MODEL_NAME,
|
|
@@ -197,10 +200,10 @@ def download_and_load_model(progress=gr.Progress()):
|
|
| 197 |
model_loaded = True
|
| 198 |
|
| 199 |
progress(1.0, desc="Model loaded successfully!")
|
| 200 |
-
return "Model loaded successfully! You can now run queries."
|
| 201 |
|
| 202 |
except Exception as e:
|
| 203 |
-
return f"Error loading model: {str(e)}
|
| 204 |
|
| 205 |
def get_tool_descriptions() -> str:
|
| 206 |
descriptions = []
|
|
@@ -455,7 +458,7 @@ with gr.Blocks(title="LLM Reasoning Modes Comparison") as demo:
|
|
| 455 |
gr.Markdown("""
|
| 456 |
# LLM Reasoning Modes Comparison
|
| 457 |
|
| 458 |
-
Compare three reasoning approaches using **
|
| 459 |
|
| 460 |
- **Think-Only**: Chain-of-Thought reasoning only (no tools)
|
| 461 |
- **Act-Only**: Tool use only (no explicit reasoning)
|
|
|
|
| 170 |
),
|
| 171 |
]
|
| 172 |
|
| 173 |
+
MODEL_NAME = "meta-llama/Llama-3.2-3B-Instruct"
|
| 174 |
model = None
|
| 175 |
tokenizer = None
|
| 176 |
model_loaded = False
|
|
|
|
| 185 |
progress(0.2, desc="Downloading tokenizer...")
|
| 186 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
|
| 187 |
|
| 188 |
+
if tokenizer.pad_token is None:
|
| 189 |
+
tokenizer.pad_token = tokenizer.eos_token
|
| 190 |
+
|
| 191 |
progress(0.4, desc="Downloading model (this may take several minutes)...")
|
| 192 |
model = AutoModelForCausalLM.from_pretrained(
|
| 193 |
MODEL_NAME,
|
|
|
|
| 200 |
model_loaded = True
|
| 201 |
|
| 202 |
progress(1.0, desc="Model loaded successfully!")
|
| 203 |
+
return f"Model '{MODEL_NAME}' loaded successfully! You can now run queries."
|
| 204 |
|
| 205 |
except Exception as e:
|
| 206 |
+
return f"Error loading model: {str(e)}"
|
| 207 |
|
| 208 |
def get_tool_descriptions() -> str:
|
| 209 |
descriptions = []
|
|
|
|
| 458 |
gr.Markdown("""
|
| 459 |
# LLM Reasoning Modes Comparison
|
| 460 |
|
| 461 |
+
Compare three reasoning approaches using **Llama-3.2-3B-Instruct**:
|
| 462 |
|
| 463 |
- **Think-Only**: Chain-of-Thought reasoning only (no tools)
|
| 464 |
- **Act-Only**: Tool use only (no explicit reasoning)
|