chichomitko commited on
Commit
1cc47b4
·
verified ·
1 Parent(s): 7b061fd

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -5
app.py CHANGED
@@ -12,12 +12,12 @@ DEFAULT_MAX_NEW_TOKENS = 1024
12
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
13
  ACCESS_TOKEN = os.getenv("HF_TOKEN", "")
14
 
15
- model_id = "bartowski/Mistral-Nemo-Instruct-2407-GGUF"
16
- filename = "Mistral-Nemo-Instruct-2407-Q6_K_L.gguf"
17
 
18
  model = AutoModelForCausalLM.from_pretrained(
19
  model_id,
20
- gguf_file=filename,
21
  torch_dtype=torch.float16,
22
  device_map="auto",
23
  trust_remote_code=False,
@@ -25,7 +25,7 @@ model = AutoModelForCausalLM.from_pretrained(
25
 
26
  tokenizer = AutoTokenizer.from_pretrained(
27
  model_id,
28
- gguf_file=filename,
29
  trust_remote_code=False,
30
  token=ACCESS_TOKEN)
31
 
@@ -109,7 +109,7 @@ chat_interface = gr.Interface(
109
  value=0.7,
110
  ),
111
  ],
112
- title="Model testing - mistralai/Mistral-Nemo-Instruct-2407",
113
  description="Provide system settings and a prompt to interact with the model.",
114
  )
115
 
 
12
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
13
  ACCESS_TOKEN = os.getenv("HF_TOKEN", "")
14
 
15
+ model_id = "nvidia/Mistral-NeMo-Minitron-8B-Instruct"
16
+ #filename = "Mistral-Nemo-Instruct-2407-Q6_K_L.gguf"
17
 
18
  model = AutoModelForCausalLM.from_pretrained(
19
  model_id,
20
+ #gguf_file=filename,
21
  torch_dtype=torch.float16,
22
  device_map="auto",
23
  trust_remote_code=False,
 
25
 
26
  tokenizer = AutoTokenizer.from_pretrained(
27
  model_id,
28
+ #gguf_file=filename,
29
  trust_remote_code=False,
30
  token=ACCESS_TOKEN)
31
 
 
109
  value=0.7,
110
  ),
111
  ],
112
+ title="Model testing - Mistral-NeMo-Minitron-8B-Instruct",
113
  description="Provide system settings and a prompt to interact with the model.",
114
  )
115