Tiberiw commited on
Commit
d7d57ec
·
1 Parent(s): 9c92b55

Update application file

Browse files
Files changed (1) hide show
  1. app.py +25 -5
app.py CHANGED
@@ -15,6 +15,13 @@ transcriber = None
15
  @asynccontextmanager
16
  async def lifespan(app: FastAPI):
17
  global transcriber
 
 
 
 
 
 
 
18
  device = "cuda:0" if torch.cuda.is_available() else "cpu"
19
  torch_dtype = torch.float16 if device == "cuda:0" else torch.float32
20
  load_dotenv(override=True) # Load environment variables from .env file
@@ -26,13 +33,26 @@ async def lifespan(app: FastAPI):
26
  if hf_token is None:
27
  raise ValueError("Hugging Face token not found. Please set the HUGGING_FACE_TOKEN environment variable.")
28
 
29
- BASE_MODEL_PATH = "openai/whisper-base"
30
  # BASE_MODEL_PATH = "openai/whisper-large-v3-turbo"
31
- ADAPTER_AND_PROCESSOR_PATH = "Tiberiw/whisper-base-lora-finetuned-custom-v1"
32
  # ADAPTER_AND_PROCESSOR_PATH = "Tiberiw/whisper-large-turbo-lora-finetuned-v3"
33
- processor = WhisperProcessor.from_pretrained(ADAPTER_AND_PROCESSOR_PATH, token=hf_token)
34
- base_model = WhisperForConditionalGeneration.from_pretrained(BASE_MODEL_PATH, torch_dtype=torch_dtype)
35
- final_model = PeftModel.from_pretrained(base_model, ADAPTER_AND_PROCESSOR_PATH, token=hf_token)
 
 
 
 
 
 
 
 
 
 
 
 
 
36
  transcriber = pipeline(
37
  "automatic-speech-recognition",
38
  model=final_model,
 
15
  @asynccontextmanager
16
  async def lifespan(app: FastAPI):
17
  global transcriber
18
+
19
+ cache_dir = os.path.join(os.getcwd(), "hf_cache")
20
+ os.makedirs(cache_dir, exist_ok=True)
21
+ os.environ["HF_HOME"] = cache_dir
22
+ os.environ["TRANSFORMERS_CACHE"] = cache_dir
23
+ os.environ["HF_HUB_CACHE"] = cache_dir
24
+
25
  device = "cuda:0" if torch.cuda.is_available() else "cpu"
26
  torch_dtype = torch.float16 if device == "cuda:0" else torch.float32
27
  load_dotenv(override=True) # Load environment variables from .env file
 
33
  if hf_token is None:
34
  raise ValueError("Hugging Face token not found. Please set the HUGGING_FACE_TOKEN environment variable.")
35
 
36
+ BASE_MODEL_PATH = "openai/whisper-large-v3-turbo"
37
  # BASE_MODEL_PATH = "openai/whisper-large-v3-turbo"
38
+ ADAPTER_AND_PROCESSOR_PATH = "Tiberiw/whisper-large-turbo-lora-finetuned-v3"
39
  # ADAPTER_AND_PROCESSOR_PATH = "Tiberiw/whisper-large-turbo-lora-finetuned-v3"
40
+ processor = WhisperProcessor.from_pretrained(
41
+ ADAPTER_AND_PROCESSOR_PATH,
42
+ token=hf_token,
43
+ cache_dir=cache_dir
44
+ )
45
+ base_model = WhisperForConditionalGeneration.from_pretrained(
46
+ BASE_MODEL_PATH,
47
+ torch_dtype=torch_dtype,
48
+ cache_dir=cache_dir
49
+ )
50
+ final_model = PeftModel.from_pretrained(
51
+ base_model,
52
+ ADAPTER_AND_PROCESSOR_PATH,
53
+ token=hf_token,
54
+ cache_dir=cache_dir
55
+ )
56
  transcriber = pipeline(
57
  "automatic-speech-recognition",
58
  model=final_model,