Update app.py
Browse files
app.py
CHANGED
|
@@ -88,6 +88,9 @@ class EventScraper:
|
|
| 88 |
|
| 89 |
def generate_with_model(self, prompt):
|
| 90 |
"""Generate text using either local model or inference client"""
|
|
|
|
|
|
|
|
|
|
| 91 |
if self.model and self.tokenizer:
|
| 92 |
# Use local model
|
| 93 |
inputs = self.tokenizer(prompt, return_tensors="pt").to(self.model.device)
|
|
@@ -236,6 +239,10 @@ class EventScraper:
|
|
| 236 |
|
| 237 |
# Generate response
|
| 238 |
response = self.generate_with_model(prompt)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 239 |
|
| 240 |
# Parse events
|
| 241 |
parsed_events = self.parse_llm_response(response)
|
|
|
|
| 88 |
|
| 89 |
def generate_with_model(self, prompt):
|
| 90 |
"""Generate text using either local model or inference client"""
|
| 91 |
+
print("------ PROMPT ------------")
|
| 92 |
+
print(prompt)
|
| 93 |
+
print("------ PROMPT ------------")
|
| 94 |
if self.model and self.tokenizer:
|
| 95 |
# Use local model
|
| 96 |
inputs = self.tokenizer(prompt, return_tensors="pt").to(self.model.device)
|
|
|
|
| 239 |
|
| 240 |
# Generate response
|
| 241 |
response = self.generate_with_model(prompt)
|
| 242 |
+
|
| 243 |
+
print("------ response ------------")
|
| 244 |
+
print(response)
|
| 245 |
+
print("------ response ------------")
|
| 246 |
|
| 247 |
# Parse events
|
| 248 |
parsed_events = self.parse_llm_response(response)
|