feyzamiynat commited on
Commit
e0ae1d8
·
verified ·
1 Parent(s): a5f0288

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +13 -2
app.py CHANGED
@@ -5,11 +5,18 @@ from transformers import AutoTokenizer, AutoModelForCausalLM, TextStreamer
5
  import logging
6
  import spaces
7
  import os
 
8
 
9
  # Logger yapılandırması
10
  logging.basicConfig(level=logging.INFO)
11
  logger = logging.getLogger(__name__)
12
 
 
 
 
 
 
 
13
  # Cihaz seçimi
14
  device = "cuda" if torch.cuda.is_available() else "cpu"
15
  logger.info(f"Kullanılan cihaz: {device}")
@@ -24,11 +31,15 @@ blip_model = BlipForConditionalGeneration.from_pretrained(
24
  # Gemma modeli yükleniyor
25
  logger.info("Gemma modeli yükleniyor...")
26
  gemma_model_id = "google/gemma-3-12b-it"
27
- gemma_tokenizer = AutoTokenizer.from_pretrained(gemma_model_id)
 
 
 
28
  gemma_model = AutoModelForCausalLM.from_pretrained(
29
  gemma_model_id,
30
  torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
31
- device_map="auto"
 
32
  )
33
  streamer = TextStreamer(gemma_tokenizer, skip_prompt=True, skip_special_tokens=True)
34
 
 
5
  import logging
6
  import spaces
7
  import os
8
+ from huggingface_hub import login
9
 
10
  # Logger yapılandırması
11
  logging.basicConfig(level=logging.INFO)
12
  logger = logging.getLogger(__name__)
13
 
14
+ # Hugging Face token kontrolü
15
+ hf_token = os.environ.get("HF_TOKEN")
16
+ if not hf_token:
17
+ raise ValueError("HF_TOKEN çevre değişkeni ayarlanmamış. Lütfen Hugging Face token'ınızı ayarlayın.")
18
+ login(token=hf_token)
19
+
20
  # Cihaz seçimi
21
  device = "cuda" if torch.cuda.is_available() else "cpu"
22
  logger.info(f"Kullanılan cihaz: {device}")
 
31
  # Gemma modeli yükleniyor
32
  logger.info("Gemma modeli yükleniyor...")
33
  gemma_model_id = "google/gemma-3-12b-it"
34
+ gemma_tokenizer = AutoTokenizer.from_pretrained(
35
+ gemma_model_id,
36
+ token=hf_token
37
+ )
38
  gemma_model = AutoModelForCausalLM.from_pretrained(
39
  gemma_model_id,
40
  torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
41
+ device_map="auto",
42
+ token=hf_token
43
  )
44
  streamer = TextStreamer(gemma_tokenizer, skip_prompt=True, skip_special_tokens=True)
45