Sarthak-506 commited on
Commit
cc98cf4
·
verified ·
1 Parent(s): 402b413

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -8
app.py CHANGED
@@ -1,6 +1,6 @@
1
  import gradio as gr
2
  from peft import PeftModel
3
- from transformers import RobertaTokenizer, T5ForConditionalGeneration, AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
4
  import torch
5
 
6
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
@@ -13,11 +13,11 @@ model_id = {
13
  }
14
 
15
  # Quantization Config
16
- quantization_config = BitsAndBytesConfig(
17
- load_in_4bit=True,
18
- bnb_4bit_quant_type="nf4",
19
- bnb_4bit_compute_dtype=torch.bfloat16
20
- )
21
 
22
  # Load CodeT5 models
23
  codeT5_tokenizer = RobertaTokenizer.from_pretrained(model_id['CodeT5'])
@@ -30,14 +30,14 @@ fine_tuned_codeT5.eval()
30
  # Load StarCoder
31
  StarCoder_tokenizer = AutoTokenizer.from_pretrained(model_id['StarCoder'])
32
  StarCoder = AutoModelForCausalLM.from_pretrained(
33
- model_id['StarCoder'], quantization_config=quantization_config, device_map="auto"
34
  )
35
  StarCoder.eval()
36
 
37
  # Load CodeLlama
38
  CodeLlama_tokenizer = AutoTokenizer.from_pretrained(model_id['CodeLlama'])
39
  CodeLlama = AutoModelForCausalLM.from_pretrained(
40
- model_id['CodeLlama'], quantization_config=quantization_config, device_map="auto"
41
  )
42
  CodeLlama.eval()
43
 
 
1
  import gradio as gr
2
  from peft import PeftModel
3
+ from transformers import RobertaTokenizer, T5ForConditionalGeneration, AutoTokenizer, AutoModelForCausalLM
4
  import torch
5
 
6
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
 
13
  }
14
 
15
  # Quantization Config
16
+ # quantization_config = BitsAndBytesConfig(
17
+ # load_in_4bit=True,
18
+ # bnb_4bit_quant_type="nf4",
19
+ # bnb_4bit_compute_dtype=torch.bfloat16
20
+ # )
21
 
22
  # Load CodeT5 models
23
  codeT5_tokenizer = RobertaTokenizer.from_pretrained(model_id['CodeT5'])
 
30
  # Load StarCoder
31
  StarCoder_tokenizer = AutoTokenizer.from_pretrained(model_id['StarCoder'])
32
  StarCoder = AutoModelForCausalLM.from_pretrained(
33
+ model_id['StarCoder'], device_map="auto"
34
  )
35
  StarCoder.eval()
36
 
37
  # Load CodeLlama
38
  CodeLlama_tokenizer = AutoTokenizer.from_pretrained(model_id['CodeLlama'])
39
  CodeLlama = AutoModelForCausalLM.from_pretrained(
40
+ model_id['CodeLlama'], device_map="auto"
41
  )
42
  CodeLlama.eval()
43