Antonio Toro Jaén commited on
Commit
f0e27e6
·
1 Parent(s): 5d94b8c
Files changed (2) hide show
  1. app.py +9 -3
  2. requirements.txt +1 -1
app.py CHANGED
@@ -7,16 +7,22 @@ import csv
7
  from huggingface_hub import login
8
 
9
 
10
- from transformers import AutoModelForCausalLM, AutoTokenizer
11
  import torch
12
 
13
  model_name = "atorojaen/DeepSeekMisongynyLyrics"
 
 
 
 
 
14
 
15
  tokenizer = AutoTokenizer.from_pretrained(model_name)
 
16
  model = AutoModelForCausalLM.from_pretrained(
17
  model_name,
18
- device_map="auto", # o "cuda" si quieres forzarlo en GPU
19
- torch_dtype=torch.float16 # si tu modelo lo soporta
20
  )
21
 
22
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
 
7
  from huggingface_hub import login
8
 
9
 
10
+ from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig
11
  import torch
12
 
13
  model_name = "atorojaen/DeepSeekMisongynyLyrics"
14
+ bnb_config = BitsAndBytesConfig(
15
+ load_in_4bit=True,
16
+ bnb_4bit_use_double_quant=True,
17
+ bnb_4bit_compute_dtype=torch.float16,
18
+ )
19
 
20
  tokenizer = AutoTokenizer.from_pretrained(model_name)
21
+
22
  model = AutoModelForCausalLM.from_pretrained(
23
  model_name,
24
+ quantization_config=bnb_config,
25
+ device_map="auto"
26
  )
27
 
28
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
requirements.txt CHANGED
@@ -3,5 +3,5 @@ torch
3
  transformers
4
  unsloth
5
  accelerate
6
- bitsandbytes
7
  scipy # requerido por algunos backends de HF
 
3
  transformers
4
  unsloth
5
  accelerate
6
+ bitsandbytes==0.42.0
7
  scipy # requerido por algunos backends de HF