Spaces:
Build error
Build error
Update services/llama_generator.py
Browse files
services/llama_generator.py
CHANGED
|
@@ -6,7 +6,9 @@ from datetime import datetime
|
|
| 6 |
import logging
|
| 7 |
from config.config import settings
|
| 8 |
|
| 9 |
-
|
|
|
|
|
|
|
| 10 |
from services.base_generator import BaseGenerator
|
| 11 |
|
| 12 |
import asyncio
|
|
@@ -46,6 +48,7 @@ class LlamaGenerator(BaseGenerator):
|
|
| 46 |
print(llama_model_name)
|
| 47 |
print(prm_model_path)
|
| 48 |
|
|
|
|
| 49 |
|
| 50 |
self.tokenizer = model_manager.load_tokenizer(llama_model_name) # Add this line to initialize the tokenizer
|
| 51 |
|
|
|
|
| 6 |
import logging
|
| 7 |
from config.config import settings
|
| 8 |
|
| 9 |
+
from services.prompt_builder import LlamaPromptTemplate
|
| 10 |
+
from services.model_manager import ModelManager
|
| 11 |
+
|
| 12 |
from services.base_generator import BaseGenerator
|
| 13 |
|
| 14 |
import asyncio
|
|
|
|
| 48 |
print(llama_model_name)
|
| 49 |
print(prm_model_path)
|
| 50 |
|
| 51 |
+
self.model_manager = ModelManager()
|
| 52 |
|
| 53 |
self.tokenizer = model_manager.load_tokenizer(llama_model_name) # Add this line to initialize the tokenizer
|
| 54 |
|