HAMMALE commited on
Commit
09b8bf6
·
verified ·
1 Parent(s): 7f75f68

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -4
app.py CHANGED
@@ -170,7 +170,7 @@ TOOLS = [
170
  ),
171
  ]
172
 
173
- MODEL_NAME = "openai/gpt-oss-20b"
174
  model = None
175
  tokenizer = None
176
  model_loaded = False
@@ -185,6 +185,9 @@ def download_and_load_model(progress=gr.Progress()):
185
  progress(0.2, desc="Downloading tokenizer...")
186
  tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
187
 
 
 
 
188
  progress(0.4, desc="Downloading model (this may take several minutes)...")
189
  model = AutoModelForCausalLM.from_pretrained(
190
  MODEL_NAME,
@@ -197,10 +200,10 @@ def download_and_load_model(progress=gr.Progress()):
197
  model_loaded = True
198
 
199
  progress(1.0, desc="Model loaded successfully!")
200
- return "Model loaded successfully! You can now run queries."
201
 
202
  except Exception as e:
203
- return f"Error loading model: {str(e)}\n\nNote: This model requires significant resources. Consider using a smaller model or API access."
204
 
205
  def get_tool_descriptions() -> str:
206
  descriptions = []
@@ -455,7 +458,7 @@ with gr.Blocks(title="LLM Reasoning Modes Comparison") as demo:
455
  gr.Markdown("""
456
  # LLM Reasoning Modes Comparison
457
 
458
- Compare three reasoning approaches using **openai/gpt-oss-20b**:
459
 
460
  - **Think-Only**: Chain-of-Thought reasoning only (no tools)
461
  - **Act-Only**: Tool use only (no explicit reasoning)
 
170
  ),
171
  ]
172
 
173
+ MODEL_NAME = "meta-llama/Llama-3.2-3B-Instruct"
174
  model = None
175
  tokenizer = None
176
  model_loaded = False
 
185
  progress(0.2, desc="Downloading tokenizer...")
186
  tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
187
 
188
+ if tokenizer.pad_token is None:
189
+ tokenizer.pad_token = tokenizer.eos_token
190
+
191
  progress(0.4, desc="Downloading model (this may take several minutes)...")
192
  model = AutoModelForCausalLM.from_pretrained(
193
  MODEL_NAME,
 
200
  model_loaded = True
201
 
202
  progress(1.0, desc="Model loaded successfully!")
203
+ return f"Model '{MODEL_NAME}' loaded successfully! You can now run queries."
204
 
205
  except Exception as e:
206
+ return f"Error loading model: {str(e)}"
207
 
208
  def get_tool_descriptions() -> str:
209
  descriptions = []
 
458
  gr.Markdown("""
459
  # LLM Reasoning Modes Comparison
460
 
461
+ Compare three reasoning approaches using **Llama-3.2-3B-Instruct**:
462
 
463
  - **Think-Only**: Chain-of-Thought reasoning only (no tools)
464
  - **Act-Only**: Tool use only (no explicit reasoning)