Chris4K commited on
Commit
3dc8018
·
verified ·
1 Parent(s): d03a443

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -0
app.py CHANGED
@@ -88,6 +88,9 @@ class EventScraper:
88
 
89
  def generate_with_model(self, prompt):
90
  """Generate text using either local model or inference client"""
 
 
 
91
  if self.model and self.tokenizer:
92
  # Use local model
93
  inputs = self.tokenizer(prompt, return_tensors="pt").to(self.model.device)
@@ -236,6 +239,10 @@ class EventScraper:
236
 
237
  # Generate response
238
  response = self.generate_with_model(prompt)
 
 
 
 
239
 
240
  # Parse events
241
  parsed_events = self.parse_llm_response(response)
 
88
 
89
  def generate_with_model(self, prompt):
90
  """Generate text using either local model or inference client"""
91
+ print("------ PROMPT ------------")
92
+ print(prompt)
93
+ print("------ PROMPT ------------")
94
  if self.model and self.tokenizer:
95
  # Use local model
96
  inputs = self.tokenizer(prompt, return_tensors="pt").to(self.model.device)
 
239
 
240
  # Generate response
241
  response = self.generate_with_model(prompt)
242
+
243
+ print("------ response ------------")
244
+ print(response)
245
+ print("------ response ------------")
246
 
247
  # Parse events
248
  parsed_events = self.parse_llm_response(response)