operablepattern commited on
Commit
8a7e475
·
verified ·
1 Parent(s): c8af86a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -2
app.py CHANGED
@@ -4,12 +4,14 @@ from llama_cpp import Llama
4
  try:
5
  llm = Llama.from_pretrained(
6
  repo_id="operablepattern/gemma-2b-it-Q",
7
- filename="*Q4_K_M.gguf",
 
8
  verbose=True
9
  )
10
  except:
11
  llm = Llama(
12
  model_path="./gemma-2b-it-Q5_K_M.gguf",
 
13
  verbose=True
14
  )
15
 
@@ -21,5 +23,4 @@ def response(message, history):
21
  gr.ChatInterface(
22
  fn=response,
23
  title="Chat with Gemma",
24
- theme='sky'
25
  ).queue().launch()
 
4
  try:
5
  llm = Llama.from_pretrained(
6
  repo_id="operablepattern/gemma-2b-it-Q",
7
+ filename="*Q5_K_M.gguf",
8
+ max_tokens=32,
9
  verbose=True
10
  )
11
  except:
12
  llm = Llama(
13
  model_path="./gemma-2b-it-Q5_K_M.gguf",
14
+ max_tokens=32,
15
  verbose=True
16
  )
17
 
 
23
  gr.ChatInterface(
24
  fn=response,
25
  title="Chat with Gemma",
 
26
  ).queue().launch()