nexusbert commited on
Commit
f28c3d1
·
1 Parent(s): b60c851
Files changed (2) hide show
  1. Dockerfile +1 -10
  2. app.py +1 -1
Dockerfile CHANGED
@@ -46,16 +46,7 @@ RUN python -c "from transformers import pipeline; pipeline('text-to-speech', mod
46
  && python -c "from transformers import pipeline; pipeline('text-to-speech', model='facebook/mms-tts-eng')" \
47
  && python -c "from transformers import pipeline; pipeline('text-to-speech', model='facebook/mms-tts-yor')"
48
 
49
- # Pre-load N-ATLaS model during build
50
- RUN python -c "import os; token = os.getenv('HF_TOKEN'); print(f'HF_TOKEN available: {bool(token)}'); print(f'Token length: {len(token) if token else 0}'); from huggingface_hub import snapshot_download; snapshot_download(repo_id='NCAIR1/N-ATLaS', token=token)" \
51
- && python -c "import os; from transformers import AutoTokenizer, AutoModelForCausalLM; import torch; token = os.getenv('HF_TOKEN'); tokenizer = AutoTokenizer.from_pretrained('NCAIR1/N-ATLaS', token=token); model = AutoModelForCausalLM.from_pretrained('NCAIR1/N-ATLaS', torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32, device_map='auto' if torch.cuda.is_available() else None, trust_remote_code=True, low_cpu_mem_usage=True, use_cache=True, token=token); print('N-ATLaS model loaded successfully')"
52
-
53
- # Pre-download ASR models (will be lazy-loaded at runtime)
54
- RUN python -c "import os; token = os.getenv('HF_TOKEN'); print(f'ASR download - HF_TOKEN available: {bool(token)}'); from huggingface_hub import snapshot_download; snapshot_download(repo_id='NCAIR1/Hausa-ASR', token=token)" \
55
- && python -c "import os; token = os.getenv('HF_TOKEN'); from huggingface_hub import snapshot_download; snapshot_download(repo_id='NCAIR1/Yoruba-ASR', token=token)" \
56
- && python -c "import os; token = os.getenv('HF_TOKEN'); from huggingface_hub import snapshot_download; snapshot_download(repo_id='NCAIR1/Igbo-ASR', token=token)" \
57
- && python -c "import os; token = os.getenv('HF_TOKEN'); from huggingface_hub import snapshot_download; snapshot_download(repo_id='NCAIR1/NigerianAccentedEnglish', token=token)" \
58
- && find /models/huggingface -name '*.lock' -delete
59
 
60
  # Copy project files
61
  COPY . .
 
46
  && python -c "from transformers import pipeline; pipeline('text-to-speech', model='facebook/mms-tts-eng')" \
47
  && python -c "from transformers import pipeline; pipeline('text-to-speech', model='facebook/mms-tts-yor')"
48
 
49
+ # NCAIR1 models will be downloaded at runtime when HF_TOKEN is available
 
 
 
 
 
 
 
 
 
50
 
51
  # Copy project files
52
  COPY . .
app.py CHANGED
@@ -214,7 +214,7 @@ def _load_natlas():
214
 
215
  try:
216
  logger.info("Loading N-ATLaS language identification model...")
217
- logger.info("Model files are pre-cached from Docker build, loading should be faster...")
218
 
219
  natlas_tokenizer = AutoTokenizer.from_pretrained("NCAIR1/N-ATLaS", token=hf_token)
220
  natlas_model = AutoModelForCausalLM.from_pretrained(
 
214
 
215
  try:
216
  logger.info("Loading N-ATLaS language identification model...")
217
+ logger.info("Downloading model files from Hugging Face...")
218
 
219
  natlas_tokenizer = AutoTokenizer.from_pretrained("NCAIR1/N-ATLaS", token=hf_token)
220
  natlas_model = AutoModelForCausalLM.from_pretrained(