jasvir-singh1021 commited on
Commit
7a6b61d
·
verified ·
1 Parent(s): bff4c38

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -3
app.py CHANGED
@@ -2,10 +2,13 @@ import gradio as gr
2
  from ctransformers import AutoModelForCausalLM
3
  import time
4
 
5
- # Load the quantized GGUF model (optimized for CPU)
 
 
 
6
  llm = AutoModelForCausalLM.from_pretrained(
7
- "TheBloke/WizardCoder-Python-13B-GGUF", # You can change to CodeLlama, Phind, etc.
8
- model_file="wizardcoder-python-13b.Q4_K_M.gguf", # Use Q4_K_M for 16GB RAM
9
  model_type="llama",
10
  config={
11
  "max_new_tokens": 512,
 
2
  from ctransformers import AutoModelForCausalLM
3
  import time
4
 
5
+ # Note: If this is a private or gated repo, you need to authenticate with your HF token.
6
+ # You can do this by setting the environment variable HF_HOME or passing token to from_pretrained.
7
+ # For public repos, no extra auth needed.
8
+
9
  llm = AutoModelForCausalLM.from_pretrained(
10
+ "TheBloke/WizardCoder-Python-7B-V1.0-GGUF", # Updated model name
11
+ model_file="wizardcoder-python-7b-v1.0.Q4_K_M.gguf", # Adjust the filename accordingly
12
  model_type="llama",
13
  config={
14
  "max_new_tokens": 512,