Spaces:
Sleeping
Sleeping
push
Browse files- Dockerfile +1 -10
- app.py +1 -1
Dockerfile
CHANGED
|
@@ -46,16 +46,7 @@ RUN python -c "from transformers import pipeline; pipeline('text-to-speech', mod
|
|
| 46 |
&& python -c "from transformers import pipeline; pipeline('text-to-speech', model='facebook/mms-tts-eng')" \
|
| 47 |
&& python -c "from transformers import pipeline; pipeline('text-to-speech', model='facebook/mms-tts-yor')"
|
| 48 |
|
| 49 |
-
#
|
| 50 |
-
RUN python -c "import os; token = os.getenv('HF_TOKEN'); print(f'HF_TOKEN available: {bool(token)}'); print(f'Token length: {len(token) if token else 0}'); from huggingface_hub import snapshot_download; snapshot_download(repo_id='NCAIR1/N-ATLaS', token=token)" \
|
| 51 |
-
&& python -c "import os; from transformers import AutoTokenizer, AutoModelForCausalLM; import torch; token = os.getenv('HF_TOKEN'); tokenizer = AutoTokenizer.from_pretrained('NCAIR1/N-ATLaS', token=token); model = AutoModelForCausalLM.from_pretrained('NCAIR1/N-ATLaS', torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32, device_map='auto' if torch.cuda.is_available() else None, trust_remote_code=True, low_cpu_mem_usage=True, use_cache=True, token=token); print('N-ATLaS model loaded successfully')"
|
| 52 |
-
|
| 53 |
-
# Pre-download ASR models (will be lazy-loaded at runtime)
|
| 54 |
-
RUN python -c "import os; token = os.getenv('HF_TOKEN'); print(f'ASR download - HF_TOKEN available: {bool(token)}'); from huggingface_hub import snapshot_download; snapshot_download(repo_id='NCAIR1/Hausa-ASR', token=token)" \
|
| 55 |
-
&& python -c "import os; token = os.getenv('HF_TOKEN'); from huggingface_hub import snapshot_download; snapshot_download(repo_id='NCAIR1/Yoruba-ASR', token=token)" \
|
| 56 |
-
&& python -c "import os; token = os.getenv('HF_TOKEN'); from huggingface_hub import snapshot_download; snapshot_download(repo_id='NCAIR1/Igbo-ASR', token=token)" \
|
| 57 |
-
&& python -c "import os; token = os.getenv('HF_TOKEN'); from huggingface_hub import snapshot_download; snapshot_download(repo_id='NCAIR1/NigerianAccentedEnglish', token=token)" \
|
| 58 |
-
&& find /models/huggingface -name '*.lock' -delete
|
| 59 |
|
| 60 |
# Copy project files
|
| 61 |
COPY . .
|
|
|
|
| 46 |
&& python -c "from transformers import pipeline; pipeline('text-to-speech', model='facebook/mms-tts-eng')" \
|
| 47 |
&& python -c "from transformers import pipeline; pipeline('text-to-speech', model='facebook/mms-tts-yor')"
|
| 48 |
|
| 49 |
+
# NCAIR1 models will be downloaded at runtime when HF_TOKEN is available
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 50 |
|
| 51 |
# Copy project files
|
| 52 |
COPY . .
|
app.py
CHANGED
|
@@ -214,7 +214,7 @@ def _load_natlas():
|
|
| 214 |
|
| 215 |
try:
|
| 216 |
logger.info("Loading N-ATLaS language identification model...")
|
| 217 |
-
logger.info("
|
| 218 |
|
| 219 |
natlas_tokenizer = AutoTokenizer.from_pretrained("NCAIR1/N-ATLaS", token=hf_token)
|
| 220 |
natlas_model = AutoModelForCausalLM.from_pretrained(
|
|
|
|
| 214 |
|
| 215 |
try:
|
| 216 |
logger.info("Loading N-ATLaS language identification model...")
|
| 217 |
+
logger.info("Downloading model files from Hugging Face...")
|
| 218 |
|
| 219 |
natlas_tokenizer = AutoTokenizer.from_pretrained("NCAIR1/N-ATLaS", token=hf_token)
|
| 220 |
natlas_model = AutoModelForCausalLM.from_pretrained(
|