runtime error
Exit code: 1. Reason: ode. To avoid downloading new versions of the code file, you can pin a revision. A new version of the following files was downloaded from https://huggingface.co/microsoft/Phi-3-mini-4k-instruct: - modeling_phi3.py . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. `flash-attention` package not found, consider installing for better performance: No module named 'flash_attn'. Current `flash-attention` does not support `window_size`. Either upgrade or use `attn_implementation='eager'`. Downloading shards: 0%| | 0/2 [00:00<?, ?it/s][A Downloading shards: 50%|█████ | 1/2 [00:15<00:15, 15.11s/it][A Downloading shards: 100%|██████████| 2/2 [00:23<00:00, 11.43s/it][A Downloading shards: 100%|██████████| 2/2 [00:23<00:00, 11.98s/it] Loading checkpoint shards: 0%| | 0/2 [00:00<?, ?it/s][A Loading checkpoint shards: 50%|█████ | 1/2 [00:06<00:06, 6.05s/it][A Loading checkpoint shards: 100%|██████████| 2/2 [00:10<00:00, 5.38s/it][A Loading checkpoint shards: 100%|██████████| 2/2 [00:10<00:00, 5.49s/it] Traceback (most recent call last): File "/home/user/app/app.py", line 5, in <module> pipe = pipeline("text-generation", model="microsoft/Phi-3-mini-4k-instruct", trust_remote_code=True, device=0) # Use device=0 for GPU, or omit for CPU File "/usr/local/lib/python3.10/site-packages/transformers/pipelines/__init__.py", line 1108, in pipeline return pipeline_class(model=model, framework=framework, task=task, **kwargs) File "/usr/local/lib/python3.10/site-packages/transformers/pipelines/text_generation.py", line 84, in __init__ super().__init__(*args, **kwargs) File "/usr/local/lib/python3.10/site-packages/transformers/pipelines/base.py", line 869, in __init__ raise ValueError(f"{device} unrecognized or not available.") ValueError: 0 unrecognized or not available.
Container logs:
Fetching error logs...