rubentsui commited on
Commit
45336c4
·
verified ·
1 Parent(s): 2c122b1

Update src/alignGenericGGUF.py

Browse files
Files changed (1) hide show
  1. src/alignGenericGGUF.py +2 -9
src/alignGenericGGUF.py CHANGED
@@ -34,14 +34,7 @@ from llama_cpp import Llama
34
  #%%
35
  start_time = datetime.now()
36
 
37
- if os.name == 'nt': # assume Windows
38
- d = 1
39
- elif os.name == 'posix':
40
- if torch.cuda.is_available(): # assume Linux
41
- d = 0
42
- elif torch.mps.is_available(): # assume macOS
43
- d = 1
44
- dev = ['cuda', 'mps', 'cpu'][d]
45
  if dev in ['cuda', 'mps']:
46
  n_gpu_layers = -1
47
  else:
@@ -49,7 +42,7 @@ else:
49
 
50
  m = 2
51
  model_name = ['Alibaba-NLP/gte-multilingual-base', 'ibm-granite/granite-embedding-278m-multilingual', 'LaBSEq80', 'LaBSEfp16', 'google/embeddinggemma-300m', 'paraphrase-multilingual-MiniLM-L12-v2'][m]
52
- model_path = ['', '', '/mnt/d/NLP/koboldcpp/ChristianAzinn/labse-gguf/labse.Q8_0.gguf', '/mnt/d/NLP/koboldcpp/ChristianAzinn/labse-gguf/labse_fp16.gguf', '', ''][m]
53
  model_name_short = ['alibaba-gte-multilingual', 'ibm-granite', 'LaBSE-gguf-q80', 'LaBSE-gguf-fp16', 'embeddinggemma-300m','paraphrase'][m]
54
  #%%
55
  print(f"Now running bitext mining with transformer model [{model_path}] on device [{dev}]...", flush=True)
 
34
  #%%
35
  start_time = datetime.now()
36
 
37
+ dev = ['cuda', 'mps', 'cpu'][2] # cpu only
 
 
 
 
 
 
 
38
  if dev in ['cuda', 'mps']:
39
  n_gpu_layers = -1
40
  else:
 
42
 
43
  m = 2
44
  model_name = ['Alibaba-NLP/gte-multilingual-base', 'ibm-granite/granite-embedding-278m-multilingual', 'LaBSEq80', 'LaBSEfp16', 'google/embeddinggemma-300m', 'paraphrase-multilingual-MiniLM-L12-v2'][m]
45
+ model_path = ['', '', 'src/labse.Q8_0.gguf', '', '', ''][m]
46
  model_name_short = ['alibaba-gte-multilingual', 'ibm-granite', 'LaBSE-gguf-q80', 'LaBSE-gguf-fp16', 'embeddinggemma-300m','paraphrase'][m]
47
  #%%
48
  print(f"Now running bitext mining with transformer model [{model_path}] on device [{dev}]...", flush=True)