aadya1762 commited on
Commit
d99243b
·
1 Parent(s): bc54c1b

bug fixes

Browse files
Files changed (2) hide show
  1. gemmademo/_chat.py +1 -2
  2. gemmademo/_model.py +5 -5
gemmademo/_chat.py CHANGED
@@ -21,8 +21,7 @@ class GradioChat:
21
  self.current_model_name = "gemma-3b"
22
  self.current_task_name = "Question Answering"
23
 
24
- # Load model lazily on first use instead of at initialization
25
- self.model = None
26
  self.prompt_manager = self._load_task(self.current_task_name)
27
 
28
  # Cache.
 
21
  self.current_model_name = "gemma-3b"
22
  self.current_task_name = "Question Answering"
23
 
24
+ self.model = self._load_model("gemma-3b")
 
25
  self.prompt_manager = self._load_task(self.current_task_name)
26
 
27
  # Cache.
gemmademo/_model.py CHANGED
@@ -51,11 +51,11 @@ class LlamaCppGemmaModel:
51
  self.messages = []
52
 
53
  # Model response generation attributes
54
- self.max_tokens = (512,)
55
- self.temperature = (0.7,)
56
- self.top_p = (0.95,)
57
- self.top_k = (40,)
58
- self.repeat_penalty = (1.1,)
59
 
60
  def load_model(self, n_ctx: int = 2048, n_gpu_layers: int = 0, system_prompt=""):
61
  """
 
51
  self.messages = []
52
 
53
  # Model response generation attributes
54
+ self.max_tokens = 512
55
+ self.temperature = 0.7
56
+ self.top_p = 0.95
57
+ self.top_k = 40
58
+ self.repeat_penalty = 1.1
59
 
60
  def load_model(self, n_ctx: int = 2048, n_gpu_layers: int = 0, system_prompt=""):
61
  """