| # Troubleshooting |
|
|
| Real failures we've hit deploying this repo to Hugging Face Inference Endpoints, and how to fix them. Read this first when the endpoint won't start. |
|
|
| --- |
|
|
| ## 1. `Unrecognized model ... Should have a model_type key in its config.json` |
| |
| Endpoint logs end with a giant list of model types (`albert, align, ... m2m_100, ... zoedepth`) and `Application startup failed`. |
|
|
| **Cause.** The Hub repo doesn't actually contain model weights / `config.json`. Usually happens when `model_loader.py` was committed to git but never *executed* against the Hub (pushing the Python file ≠ running it). |
|
|
| **Check.** |
|
|
| ```bash |
| python3 -c "from huggingface_hub import HfApi; print([s.rfilename for s in HfApi().model_info('ericaRC/example').siblings])" |
| ``` |
|
|
| You should see `config.json`, `model.safetensors`, `tokenizer_config.json`, `tokenizer.json`, `handler.py`, `requirements.txt`, `README.md`. If it's only `.gitattributes` and scripts, the weights were never pushed. |
|
|
| **Fix.** |
|
|
| ```bash |
| huggingface-cli login |
| python3 model_loader.py |
| ``` |
|
|
| --- |
|
|
| ## 2. `403 Forbidden` on `.../info/lfs/objects/batch` |
|
|
| `push_to_hub` dies with `HfHubHTTPError: 403 Forbidden: Authorization error.` |
|
|
| **Cause.** Your HF token lacks write access to the target repo. Most commonly: a fine-grained token scoped to your user only, trying to push to an org namespace. Reading works (which is why `whoami` succeeds) but LFS writes are rejected. |
|
|
| **Check.** |
|
|
| ```bash |
| python3 -c " |
| from huggingface_hub import HfApi |
| perms = HfApi().whoami()['auth']['accessToken'].get('fineGrained', {}) |
| for s in perms.get('scoped', []): |
| print(s['entity']['type'], s['entity']['name'], '->', s['permissions']) |
| " |
| ``` |
|
|
| You need an entry matching the target repo's namespace (user or org) that includes `repo.write`. |
|
|
| **Fix.** At https://huggingface.co/settings/tokens either: |
| - Edit the existing token and add the org with `repo.write` + `repo.content.read` + `repo.access.read`, **or** |
| - Create a new classic "Write" token and `huggingface-cli login` with it. |
|
|
| --- |
|
|
| ## 3. `AttributeError: 'list' object has no attribute 'keys'` in `_set_model_specific_special_tokens` |
| |
| Endpoint logs show a traceback through `tokenization_nllb_fast.py` → `tokenization_utils_base.py` and crash on: |
| |
| ``` |
| self.SPECIAL_TOKENS_ATTRIBUTES + list(special_tokens.keys()) |
| ``` |
| |
| **Cause.** Transformers-version skew between save time and load time. `transformers` 5.x introduced an `extra_special_tokens` field (serialized as a list for NLLB's Flores-200 codes). The Inference Endpoints base image ships a `transformers` 4.x that expects `extra_special_tokens` to be a dict and calls `.keys()` on it. |
| |
| **Check.** |
| |
| ```bash |
| python3 -c " |
| import json |
| from huggingface_hub import hf_hub_download |
| cfg = json.load(open(hf_hub_download('ericaRC/example', 'tokenizer_config.json'))) |
| print('extra_special_tokens type:', type(cfg.get('extra_special_tokens')).__name__) |
| print('additional_special_tokens count:', len(cfg.get('additional_special_tokens') or [])) |
| " |
| ``` |
| |
| If `extra_special_tokens` is a non-empty `list` and `additional_special_tokens` is empty, you're hitting this. |
| |
| **Fix (already applied to this repo).** `tokenizer_config.json` has been normalized: |
| - lang codes live in `additional_special_tokens` (list — old *and* new transformers accept this) |
| - `extra_special_tokens` is `{}` (empty dict — passes `.keys()` in old transformers, ignored in new) |
| |
| And `requirements.txt` pins `transformers>=4.40.0,<5.0` to prevent the endpoint from auto-pulling a 5.x that re-introduces the mismatch. |
| |
| **Prevention going forward.** When running `model_loader.py`, use the same `transformers` major version the endpoint runs: |
| |
| ```bash |
| pip install "transformers<5" "huggingface_hub" "torch" |
| python3 model_loader.py |
| ``` |
| |
| Don't save tokenizers from `transformers` 5.x and load them in a 4.x container (or vice versa) unless you've confirmed the schema matches. |
| |
| --- |
| |
| ## 4. Endpoint boots but requests return garbage / wrong language |
| |
| **Cause.** `src_lang` wasn't set on the tokenizer, or `forced_bos_token_id` wasn't passed at generation time. NLLB needs both. |
| |
| **Check.** Look at the request body: |
| |
| ```json |
| { |
| "inputs": "Hello, world!", |
| "parameters": { "src_lang": "eng_Latn", "tgt_lang": "fra_Latn" } |
| } |
| ``` |
| |
| If you're hitting the endpoint without a `parameters` block, `handler.py` falls back to `eng_Latn → spa_Latn`. |
| |
| **Fix.** Always pass `src_lang` and `tgt_lang` using [Flores-200 codes](https://github.com/facebookresearch/flores/blob/main/flores200/README.md#languages-in-flores-200). |
| |
| --- |
| |
| ## 5. Container Type is set to "Text Generation Inference (TGI)" |
| |
| TGI only supports decoder-only causal LMs. NLLB is seq2seq, so TGI will refuse to load it and `handler.py` will be ignored. |
| |
| **Fix.** In the endpoint's Advanced configuration, set **Container Type → Default** (the HF inference toolkit). That container picks up `handler.py` automatically. |
| |
| --- |
| |
| ## Checklist before clicking Deploy |
| |
| - [ ] `HfApi().model_info(REPO).siblings` lists `config.json`, `model.safetensors`, `tokenizer*.json`, `handler.py`, `requirements.txt`, `README.md`. |
| - [ ] `tokenizer_config.json` has `extra_special_tokens: {}` (or absent) and `additional_special_tokens` populated. |
| - [ ] `requirements.txt` pins `transformers<5`. |
| - [ ] Local smoke test passes: |
| ```python |
| from handler import EndpointHandler |
| h = EndpointHandler("ericaRC/example") |
| print(h({"inputs": "Hello, world!", "parameters": {"src_lang": "eng_Latn", "tgt_lang": "fra_Latn"}})) |
| ``` |
| - [ ] Endpoint Container Type = **Default**, not TGI. |
| |