Spaces:
Build error
Build error
Update app.py
Browse files
app.py
CHANGED
|
@@ -388,7 +388,7 @@ class LlamaGenerator(BaseGenerator):
|
|
| 388 |
)
|
| 389 |
|
| 390 |
# Initialize models
|
| 391 |
-
self.model_manager.load_model(
|
| 392 |
"llama",
|
| 393 |
llama_model_name,
|
| 394 |
"llama",
|
|
@@ -448,6 +448,52 @@ class LlamaGenerator(BaseGenerator):
|
|
| 448 |
model_kwargs,
|
| 449 |
**kwargs
|
| 450 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 451 |
|
| 452 |
def check_health(self) -> HealthStatus:
|
| 453 |
"""Check the health status of the generator."""
|
|
|
|
| 388 |
)
|
| 389 |
|
| 390 |
# Initialize models
|
| 391 |
+
self.model_manager.load_model(
|
| 392 |
"llama",
|
| 393 |
llama_model_name,
|
| 394 |
"llama",
|
|
|
|
| 448 |
model_kwargs,
|
| 449 |
**kwargs
|
| 450 |
)
|
| 451 |
+
|
| 452 |
+
|
| 453 |
+
def generate_with_context(
|
| 454 |
+
self,
|
| 455 |
+
context: str,
|
| 456 |
+
user_input: str,
|
| 457 |
+
chat_history: List[Tuple[str, str]],
|
| 458 |
+
model_kwargs: Dict[str, Any],
|
| 459 |
+
max_history_turns: int = 3,
|
| 460 |
+
strategy: str = "default",
|
| 461 |
+
num_samples: int = 5,
|
| 462 |
+
depth: int = 3,
|
| 463 |
+
breadth: int = 2,
|
| 464 |
+
|
| 465 |
+
) -> str:
|
| 466 |
+
"""Generate a response using context and chat history.
|
| 467 |
+
|
| 468 |
+
Args:
|
| 469 |
+
context (str): Context for the conversation
|
| 470 |
+
user_input (str): Current user input
|
| 471 |
+
chat_history (List[Tuple[str, str]]): List of (user, assistant) message pairs
|
| 472 |
+
model_kwargs (dict): Additional arguments for model.generate()
|
| 473 |
+
max_history_turns (int): Maximum number of history turns to include
|
| 474 |
+
strategy (str): Generation strategy
|
| 475 |
+
num_samples (int): Number of samples for applicable strategies
|
| 476 |
+
depth (int): Depth for DVTS strategy
|
| 477 |
+
breadth (int): Breadth for DVTS strategy
|
| 478 |
+
|
| 479 |
+
Returns:
|
| 480 |
+
str: Generated response
|
| 481 |
+
"""
|
| 482 |
+
prompt = self._construct_prompt(
|
| 483 |
+
context,
|
| 484 |
+
user_input,
|
| 485 |
+
chat_history,
|
| 486 |
+
max_history_turns
|
| 487 |
+
)
|
| 488 |
+
return self.generate(
|
| 489 |
+
prompt,
|
| 490 |
+
model_kwargs,
|
| 491 |
+
strategy,
|
| 492 |
+
num_samples,
|
| 493 |
+
depth,
|
| 494 |
+
breadth
|
| 495 |
+
)
|
| 496 |
+
|
| 497 |
|
| 498 |
def check_health(self) -> HealthStatus:
|
| 499 |
"""Check the health status of the generator."""
|