Scheduled Commit
Browse files- vllm_0004000.log +143 -20
- vllm_0005000.log +241 -32
- vllm_0006000.log +18 -20
- vllm_0006500.log +203 -97
- vllm_0007000.log +18 -20
vllm_0004000.log
CHANGED
|
@@ -1,20 +1,143 @@
|
|
| 1 |
-
[0;36m(APIServer pid=
|
| 2 |
-
[0;36m(APIServer pid=
|
| 3 |
-
[0;36m(APIServer pid=
|
| 4 |
-
[0;36m(APIServer pid=
|
| 5 |
-
[0;36m(APIServer pid=
|
| 6 |
-
[0;36m(APIServer pid=
|
| 7 |
-
[0;36m(APIServer pid=
|
| 8 |
-
[0;36m(APIServer pid=
|
| 9 |
-
[0;36m(APIServer pid=
|
| 10 |
-
[0;36m(APIServer pid=
|
| 11 |
-
[0;36m(APIServer pid=
|
| 12 |
-
[0;36m(APIServer pid=
|
| 13 |
-
[0;36m(APIServer pid=
|
| 14 |
-
[0;36m(EngineCore_DP0 pid=
|
| 15 |
-
[0;36m(EngineCore_DP0 pid=
|
| 16 |
-
[0;36m(EngineCore_DP0 pid=
|
| 17 |
-
[0;36m(EngineCore_DP0 pid=
|
| 18 |
-
[0;36m(EngineCore_DP0 pid=
|
| 19 |
-
[0;36m(EngineCore_DP0 pid=
|
| 20 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[0;36m(APIServer pid=3309550)[0;0m INFO 02-03 01:40:54 [utils.py:325]
|
| 2 |
+
[0;36m(APIServer pid=3309550)[0;0m INFO 02-03 01:40:54 [utils.py:325] β β ββ ββ
|
| 3 |
+
[0;36m(APIServer pid=3309550)[0;0m INFO 02-03 01:40:54 [utils.py:325] ββ ββ β β β βββ β version 0.15.0
|
| 4 |
+
[0;36m(APIServer pid=3309550)[0;0m INFO 02-03 01:40:54 [utils.py:325] ββββ β β β β model Elfsong/VLM_stage_2_iter_0004000
|
| 5 |
+
[0;36m(APIServer pid=3309550)[0;0m INFO 02-03 01:40:54 [utils.py:325] ββ βββββ βββββ β β
|
| 6 |
+
[0;36m(APIServer pid=3309550)[0;0m INFO 02-03 01:40:54 [utils.py:325]
|
| 7 |
+
[0;36m(APIServer pid=3309550)[0;0m INFO 02-03 01:40:54 [utils.py:261] non-default args: {'port': 9000, 'model': 'Elfsong/VLM_stage_2_iter_0004000', 'trust_remote_code': True, 'gpu_memory_utilization': 0.4}
|
| 8 |
+
[0;36m(APIServer pid=3309550)[0;0m The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored.
|
| 9 |
+
[0;36m(APIServer pid=3309550)[0;0m The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored.
|
| 10 |
+
[0;36m(APIServer pid=3309550)[0;0m INFO 02-03 01:40:56 [model.py:541] Resolved architecture: Qwen3ForCausalLM
|
| 11 |
+
[0;36m(APIServer pid=3309550)[0;0m INFO 02-03 01:40:56 [model.py:1561] Using max model len 40960
|
| 12 |
+
[0;36m(APIServer pid=3309550)[0;0m INFO 02-03 01:40:56 [scheduler.py:226] Chunked prefill is enabled with max_num_batched_tokens=8192.
|
| 13 |
+
[0;36m(APIServer pid=3309550)[0;0m INFO 02-03 01:40:56 [vllm.py:624] Asynchronous scheduling is enabled.
|
| 14 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m INFO 02-03 01:41:06 [core.py:96] Initializing a V1 LLM engine (v0.15.0) with config: model='Elfsong/VLM_stage_2_iter_0004000', speculative_config=None, tokenizer='Elfsong/VLM_stage_2_iter_0004000', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=40960, download_dir=None, load_format=auto, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, enable_return_routed_experts=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser='', reasoning_parser_plugin='', enable_in_reasoning=False), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None, kv_cache_metrics=False, kv_cache_metrics_sample=0.01, cudagraph_metrics=False, enable_layerwise_nvtx_tracing=False, enable_mfu_metrics=False, enable_mm_processor_stats=False, enable_logging_iteration_details=False), seed=0, served_model_name=Elfsong/VLM_stage_2_iter_0004000, enable_prefix_caching=True, enable_chunked_prefill=True, pooler_config=None, compilation_config={'level': None, 'mode': <CompilationMode.VLLM_COMPILE: 3>, 'debug_dump_path': None, 'cache_dir': '', 'compile_cache_save_format': 'binary', 'backend': 'inductor', 'custom_ops': ['none'], 'splitting_ops': ['vllm::unified_attention', 'vllm::unified_attention_with_output', 'vllm::unified_mla_attention', 'vllm::unified_mla_attention_with_output', 'vllm::mamba_mixer2', 'vllm::mamba_mixer', 'vllm::short_conv', 'vllm::linear_attention', 'vllm::plamo2_mamba_mixer', 'vllm::gdn_attention_core', 'vllm::kda_attention', 'vllm::sparse_attn_indexer', 'vllm::rocm_aiter_sparse_attn_indexer'], 'compile_mm_encoder': False, 'compile_sizes': [], 'compile_ranges_split_points': [8192], 'inductor_compile_config': {'enable_auto_functionalized_v2': False, 'combo_kernels': True, 'benchmark_combo_kernel': True}, 'inductor_passes': {}, 'cudagraph_mode': <CUDAGraphMode.FULL_AND_PIECEWISE: (2, 1)>, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [1, 2, 4, 8, 16, 24, 32, 40, 48, 56, 64, 72, 80, 88, 96, 104, 112, 120, 128, 136, 144, 152, 160, 168, 176, 184, 192, 200, 208, 216, 224, 232, 240, 248, 256, 272, 288, 304, 320, 336, 352, 368, 384, 400, 416, 432, 448, 464, 480, 496, 512], 'cudagraph_copy_inputs': False, 'cudagraph_specialize_lora': True, 'use_inductor_graph_partition': False, 'pass_config': {'fuse_norm_quant': False, 'fuse_act_quant': False, 'fuse_attn_quant': False, 'eliminate_noops': True, 'enable_sp': False, 'fuse_gemm_comms': False, 'fuse_allreduce_rms': False}, 'max_cudagraph_capture_size': 512, 'dynamic_shapes_config': {'type': <DynamicShapesType.BACKED: 'backed'>, 'evaluate_guards': False, 'assume_32_bit_indexing': True}, 'local_cache_dir': None}
|
| 15 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m INFO 02-03 01:41:11 [parallel_state.py:1212] world_size=1 rank=0 local_rank=0 distributed_init_method=tcp://172.21.25.98:59653 backend=nccl
|
| 16 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m INFO 02-03 01:41:11 [parallel_state.py:1423] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, PCP rank 0, TP rank 0, EP rank N/A
|
| 17 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m INFO 02-03 01:41:12 [gpu_model_runner.py:4021] Starting to load model Elfsong/VLM_stage_2_iter_0004000...
|
| 18 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m INFO 02-03 01:41:13 [cuda.py:364] Using FLASH_ATTN attention backend out of potential backends: ('FLASH_ATTN', 'FLASHINFER', 'TRITON_ATTN', 'FLEX_ATTENTION')
|
| 19 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m INFO 02-03 01:42:48 [weight_utils.py:527] Time spent downloading weights for Elfsong/VLM_stage_2_iter_0004000: 92.466791 seconds
|
| 20 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 21 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 22 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 23 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 24 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 25 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 26 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 27 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 28 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 29 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 30 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 31 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 32 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 33 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 34 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 35 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 36 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 37 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 38 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 39 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m
|
| 40 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m INFO 02-03 01:43:08 [default_loader.py:291] Loading weights took 20.23 seconds
|
| 41 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m INFO 02-03 01:43:09 [gpu_model_runner.py:4118] Model loading took 61.03 GiB memory and 115.770729 seconds
|
| 42 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m INFO 02-03 01:43:25 [backends.py:805] Using cache directory: /home/mingzhe/.cache/vllm/torch_compile_cache/226ddebc06/rank_0_0/backbone for vLLM's torch.compile
|
| 43 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m INFO 02-03 01:43:25 [backends.py:865] Dynamo bytecode transform time: 15.42 s
|
| 44 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m INFO 02-03 01:43:38 [backends.py:302] Cache the graph of compile range (1, 8192) for later use
|
| 45 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m INFO 02-03 01:44:12 [backends.py:319] Compiling a graph for compile range (1, 8192) takes 34.69 s
|
| 46 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m INFO 02-03 01:44:12 [monitor.py:34] torch.compile takes 50.11 s in total
|
| 47 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m INFO 02-03 01:44:15 [gpu_worker.py:356] Available KV cache memory: -77.05 GiB
|
| 48 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] EngineCore failed to start.
|
| 49 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] Traceback (most recent call last):
|
| 50 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 937, in run_engine_core
|
| 51 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] engine_core = EngineCoreProc(*args, engine_index=dp_rank, **kwargs)
|
| 52 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 53 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 691, in __init__
|
| 54 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] super().__init__(
|
| 55 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 112, in __init__
|
| 56 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] num_gpu_blocks, num_cpu_blocks, kv_cache_config = self._initialize_kv_caches(
|
| 57 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 58 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 253, in _initialize_kv_caches
|
| 59 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] kv_cache_configs = get_kv_cache_configs(
|
| 60 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^
|
| 61 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/core/kv_cache_utils.py", line 1516, in get_kv_cache_configs
|
| 62 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] _check_enough_kv_cache_memory(
|
| 63 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/core/kv_cache_utils.py", line 616, in _check_enough_kv_cache_memory
|
| 64 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] raise ValueError(
|
| 65 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ERROR 02-03 01:44:15 [core.py:946] ValueError: No available memory for the cache blocks. Try increasing `gpu_memory_utilization` when initializing the engine. See https://docs.vllm.ai/en/latest/configuration/conserving_memory/ for more details.
|
| 66 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m Process EngineCore_DP0:
|
| 67 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m Traceback (most recent call last):
|
| 68 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m File "/usr/lib/python3.12/multiprocessing/process.py", line 314, in _bootstrap
|
| 69 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m self.run()
|
| 70 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m File "/usr/lib/python3.12/multiprocessing/process.py", line 108, in run
|
| 71 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m self._target(*self._args, **self._kwargs)
|
| 72 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 950, in run_engine_core
|
| 73 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m raise e
|
| 74 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 937, in run_engine_core
|
| 75 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m engine_core = EngineCoreProc(*args, engine_index=dp_rank, **kwargs)
|
| 76 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 77 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 691, in __init__
|
| 78 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m super().__init__(
|
| 79 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 112, in __init__
|
| 80 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m num_gpu_blocks, num_cpu_blocks, kv_cache_config = self._initialize_kv_caches(
|
| 81 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 82 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 253, in _initialize_kv_caches
|
| 83 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m kv_cache_configs = get_kv_cache_configs(
|
| 84 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ^^^^^^^^^^^^^^^^^^^^^
|
| 85 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/core/kv_cache_utils.py", line 1516, in get_kv_cache_configs
|
| 86 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m _check_enough_kv_cache_memory(
|
| 87 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/core/kv_cache_utils.py", line 616, in _check_enough_kv_cache_memory
|
| 88 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m raise ValueError(
|
| 89 |
+
[0;36m(EngineCore_DP0 pid=3310619)[0;0m ValueError: No available memory for the cache blocks. Try increasing `gpu_memory_utilization` when initializing the engine. See https://docs.vllm.ai/en/latest/configuration/conserving_memory/ for more details.
|
| 90 |
+
[rank0]:[W203 01:44:17.466159054 ProcessGroupNCCL.cpp:1524] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
|
| 91 |
+
[0;36m(APIServer pid=3309550)[0;0m Traceback (most recent call last):
|
| 92 |
+
[0;36m(APIServer pid=3309550)[0;0m File "<frozen runpy>", line 198, in _run_module_as_main
|
| 93 |
+
[0;36m(APIServer pid=3309550)[0;0m File "<frozen runpy>", line 88, in _run_code
|
| 94 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 991, in <module>
|
| 95 |
+
[0;36m(APIServer pid=3309550)[0;0m uvloop.run(run_server(args))
|
| 96 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/uvloop/__init__.py", line 96, in run
|
| 97 |
+
[0;36m(APIServer pid=3309550)[0;0m return __asyncio.run(
|
| 98 |
+
[0;36m(APIServer pid=3309550)[0;0m ^^^^^^^^^^^^^^
|
| 99 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/usr/lib/python3.12/asyncio/runners.py", line 194, in run
|
| 100 |
+
[0;36m(APIServer pid=3309550)[0;0m return runner.run(main)
|
| 101 |
+
[0;36m(APIServer pid=3309550)[0;0m ^^^^^^^^^^^^^^^^
|
| 102 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/usr/lib/python3.12/asyncio/runners.py", line 118, in run
|
| 103 |
+
[0;36m(APIServer pid=3309550)[0;0m return self._loop.run_until_complete(task)
|
| 104 |
+
[0;36m(APIServer pid=3309550)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 105 |
+
[0;36m(APIServer pid=3309550)[0;0m File "uvloop/loop.pyx", line 1518, in uvloop.loop.Loop.run_until_complete
|
| 106 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/uvloop/__init__.py", line 48, in wrapper
|
| 107 |
+
[0;36m(APIServer pid=3309550)[0;0m return await main
|
| 108 |
+
[0;36m(APIServer pid=3309550)[0;0m ^^^^^^^^^^
|
| 109 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 919, in run_server
|
| 110 |
+
[0;36m(APIServer pid=3309550)[0;0m await run_server_worker(listen_address, sock, args, **uvicorn_kwargs)
|
| 111 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 938, in run_server_worker
|
| 112 |
+
[0;36m(APIServer pid=3309550)[0;0m async with build_async_engine_client(
|
| 113 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/usr/lib/python3.12/contextlib.py", line 210, in __aenter__
|
| 114 |
+
[0;36m(APIServer pid=3309550)[0;0m return await anext(self.gen)
|
| 115 |
+
[0;36m(APIServer pid=3309550)[0;0m ^^^^^^^^^^^^^^^^^^^^^
|
| 116 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 147, in build_async_engine_client
|
| 117 |
+
[0;36m(APIServer pid=3309550)[0;0m async with build_async_engine_client_from_engine_args(
|
| 118 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/usr/lib/python3.12/contextlib.py", line 210, in __aenter__
|
| 119 |
+
[0;36m(APIServer pid=3309550)[0;0m return await anext(self.gen)
|
| 120 |
+
[0;36m(APIServer pid=3309550)[0;0m ^^^^^^^^^^^^^^^^^^^^^
|
| 121 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 188, in build_async_engine_client_from_engine_args
|
| 122 |
+
[0;36m(APIServer pid=3309550)[0;0m async_llm = AsyncLLM.from_vllm_config(
|
| 123 |
+
[0;36m(APIServer pid=3309550)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 124 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/async_llm.py", line 228, in from_vllm_config
|
| 125 |
+
[0;36m(APIServer pid=3309550)[0;0m return cls(
|
| 126 |
+
[0;36m(APIServer pid=3309550)[0;0m ^^^^
|
| 127 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/async_llm.py", line 155, in __init__
|
| 128 |
+
[0;36m(APIServer pid=3309550)[0;0m self.engine_core = EngineCoreClient.make_async_mp_client(
|
| 129 |
+
[0;36m(APIServer pid=3309550)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 130 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core_client.py", line 122, in make_async_mp_client
|
| 131 |
+
[0;36m(APIServer pid=3309550)[0;0m return AsyncMPClient(*client_args)
|
| 132 |
+
[0;36m(APIServer pid=3309550)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 133 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core_client.py", line 819, in __init__
|
| 134 |
+
[0;36m(APIServer pid=3309550)[0;0m super().__init__(
|
| 135 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core_client.py", line 479, in __init__
|
| 136 |
+
[0;36m(APIServer pid=3309550)[0;0m with launch_core_engines(vllm_config, executor_class, log_stats) as (
|
| 137 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/usr/lib/python3.12/contextlib.py", line 144, in __exit__
|
| 138 |
+
[0;36m(APIServer pid=3309550)[0;0m next(self.gen)
|
| 139 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/utils.py", line 933, in launch_core_engines
|
| 140 |
+
[0;36m(APIServer pid=3309550)[0;0m wait_for_engine_startup(
|
| 141 |
+
[0;36m(APIServer pid=3309550)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/utils.py", line 992, in wait_for_engine_startup
|
| 142 |
+
[0;36m(APIServer pid=3309550)[0;0m raise RuntimeError(
|
| 143 |
+
[0;36m(APIServer pid=3309550)[0;0m RuntimeError: Engine core initialization failed. See root cause above. Failed core proc(s): {}
|
vllm_0005000.log
CHANGED
|
@@ -1,32 +1,241 @@
|
|
| 1 |
-
[0;36m(APIServer pid=
|
| 2 |
-
[0;36m(APIServer pid=
|
| 3 |
-
[0;36m(APIServer pid=
|
| 4 |
-
[0;36m(APIServer pid=
|
| 5 |
-
[0;36m(APIServer pid=
|
| 6 |
-
[0;36m(APIServer pid=
|
| 7 |
-
[0;36m(APIServer pid=
|
| 8 |
-
[0;36m(APIServer pid=
|
| 9 |
-
[0;36m(APIServer pid=
|
| 10 |
-
[0;36m(APIServer pid=
|
| 11 |
-
[0;36m(APIServer pid=
|
| 12 |
-
[0;36m(APIServer pid=
|
| 13 |
-
[0;36m(APIServer pid=
|
| 14 |
-
[0;36m(EngineCore_DP0 pid=
|
| 15 |
-
[0;36m(EngineCore_DP0 pid=
|
| 16 |
-
[0;36m(EngineCore_DP0 pid=
|
| 17 |
-
[0;36m(EngineCore_DP0 pid=
|
| 18 |
-
[0;36m(EngineCore_DP0 pid=
|
| 19 |
-
[0;36m(EngineCore_DP0 pid=
|
| 20 |
-
[0;36m(EngineCore_DP0 pid=
|
| 21 |
-
[0;36m(EngineCore_DP0 pid=
|
| 22 |
-
[0;36m(EngineCore_DP0 pid=
|
| 23 |
-
[0;36m(EngineCore_DP0 pid=
|
| 24 |
-
[0;36m(EngineCore_DP0 pid=
|
| 25 |
-
[0;36m(EngineCore_DP0 pid=
|
| 26 |
-
[0;36m(EngineCore_DP0 pid=
|
| 27 |
-
[0;36m(EngineCore_DP0 pid=
|
| 28 |
-
[0;36m(EngineCore_DP0 pid=
|
| 29 |
-
[0;36m(EngineCore_DP0 pid=
|
| 30 |
-
[0;36m(EngineCore_DP0 pid=
|
| 31 |
-
[0;36m(EngineCore_DP0 pid=
|
| 32 |
-
[0;36m(EngineCore_DP0 pid=
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[0;36m(APIServer pid=3309898)[0;0m INFO 02-03 01:40:59 [utils.py:325]
|
| 2 |
+
[0;36m(APIServer pid=3309898)[0;0m INFO 02-03 01:40:59 [utils.py:325] β β ββ ββ
|
| 3 |
+
[0;36m(APIServer pid=3309898)[0;0m INFO 02-03 01:40:59 [utils.py:325] ββ ββ β β β βββ β version 0.15.0
|
| 4 |
+
[0;36m(APIServer pid=3309898)[0;0m INFO 02-03 01:40:59 [utils.py:325] ββββ β β β β model Elfsong/VLM_stage_2_iter_0005000
|
| 5 |
+
[0;36m(APIServer pid=3309898)[0;0m INFO 02-03 01:40:59 [utils.py:325] ββ βββββ βββββ β β
|
| 6 |
+
[0;36m(APIServer pid=3309898)[0;0m INFO 02-03 01:40:59 [utils.py:325]
|
| 7 |
+
[0;36m(APIServer pid=3309898)[0;0m INFO 02-03 01:40:59 [utils.py:261] non-default args: {'port': 9001, 'model': 'Elfsong/VLM_stage_2_iter_0005000', 'trust_remote_code': True, 'gpu_memory_utilization': 0.4}
|
| 8 |
+
[0;36m(APIServer pid=3309898)[0;0m The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored.
|
| 9 |
+
[0;36m(APIServer pid=3309898)[0;0m The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored.
|
| 10 |
+
[0;36m(APIServer pid=3309898)[0;0m INFO 02-03 01:41:01 [model.py:541] Resolved architecture: Qwen3ForCausalLM
|
| 11 |
+
[0;36m(APIServer pid=3309898)[0;0m INFO 02-03 01:41:01 [model.py:1561] Using max model len 40960
|
| 12 |
+
[0;36m(APIServer pid=3309898)[0;0m INFO 02-03 01:41:01 [scheduler.py:226] Chunked prefill is enabled with max_num_batched_tokens=8192.
|
| 13 |
+
[0;36m(APIServer pid=3309898)[0;0m INFO 02-03 01:41:01 [vllm.py:624] Asynchronous scheduling is enabled.
|
| 14 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m INFO 02-03 01:41:12 [core.py:96] Initializing a V1 LLM engine (v0.15.0) with config: model='Elfsong/VLM_stage_2_iter_0005000', speculative_config=None, tokenizer='Elfsong/VLM_stage_2_iter_0005000', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=40960, download_dir=None, load_format=auto, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, enable_return_routed_experts=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser='', reasoning_parser_plugin='', enable_in_reasoning=False), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None, kv_cache_metrics=False, kv_cache_metrics_sample=0.01, cudagraph_metrics=False, enable_layerwise_nvtx_tracing=False, enable_mfu_metrics=False, enable_mm_processor_stats=False, enable_logging_iteration_details=False), seed=0, served_model_name=Elfsong/VLM_stage_2_iter_0005000, enable_prefix_caching=True, enable_chunked_prefill=True, pooler_config=None, compilation_config={'level': None, 'mode': <CompilationMode.VLLM_COMPILE: 3>, 'debug_dump_path': None, 'cache_dir': '', 'compile_cache_save_format': 'binary', 'backend': 'inductor', 'custom_ops': ['none'], 'splitting_ops': ['vllm::unified_attention', 'vllm::unified_attention_with_output', 'vllm::unified_mla_attention', 'vllm::unified_mla_attention_with_output', 'vllm::mamba_mixer2', 'vllm::mamba_mixer', 'vllm::short_conv', 'vllm::linear_attention', 'vllm::plamo2_mamba_mixer', 'vllm::gdn_attention_core', 'vllm::kda_attention', 'vllm::sparse_attn_indexer', 'vllm::rocm_aiter_sparse_attn_indexer'], 'compile_mm_encoder': False, 'compile_sizes': [], 'compile_ranges_split_points': [8192], 'inductor_compile_config': {'enable_auto_functionalized_v2': False, 'combo_kernels': True, 'benchmark_combo_kernel': True}, 'inductor_passes': {}, 'cudagraph_mode': <CUDAGraphMode.FULL_AND_PIECEWISE: (2, 1)>, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [1, 2, 4, 8, 16, 24, 32, 40, 48, 56, 64, 72, 80, 88, 96, 104, 112, 120, 128, 136, 144, 152, 160, 168, 176, 184, 192, 200, 208, 216, 224, 232, 240, 248, 256, 272, 288, 304, 320, 336, 352, 368, 384, 400, 416, 432, 448, 464, 480, 496, 512], 'cudagraph_copy_inputs': False, 'cudagraph_specialize_lora': True, 'use_inductor_graph_partition': False, 'pass_config': {'fuse_norm_quant': False, 'fuse_act_quant': False, 'fuse_attn_quant': False, 'eliminate_noops': True, 'enable_sp': False, 'fuse_gemm_comms': False, 'fuse_allreduce_rms': False}, 'max_cudagraph_capture_size': 512, 'dynamic_shapes_config': {'type': <DynamicShapesType.BACKED: 'backed'>, 'evaluate_guards': False, 'assume_32_bit_indexing': True}, 'local_cache_dir': None}
|
| 15 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m INFO 02-03 01:41:16 [parallel_state.py:1212] world_size=1 rank=0 local_rank=0 distributed_init_method=tcp://172.21.25.98:53221 backend=nccl
|
| 16 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m INFO 02-03 01:41:16 [parallel_state.py:1423] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, PCP rank 0, TP rank 0, EP rank N/A
|
| 17 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m INFO 02-03 01:41:17 [gpu_model_runner.py:4021] Starting to load model Elfsong/VLM_stage_2_iter_0005000...
|
| 18 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m INFO 02-03 01:41:18 [cuda.py:364] Using FLASH_ATTN attention backend out of potential backends: ('FLASH_ATTN', 'FLASHINFER', 'TRITON_ATTN', 'FLEX_ATTENTION')
|
| 19 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 20 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 21 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 22 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 23 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 24 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 25 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 26 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 27 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 28 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 29 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 30 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 31 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 32 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 33 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 34 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 35 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 36 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 37 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 38 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 39 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m INFO 02-03 01:41:42 [default_loader.py:291] Loading weights took 21.25 seconds
|
| 40 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m INFO 02-03 01:41:43 [gpu_model_runner.py:4118] Model loading took 61.03 GiB memory and 24.133907 seconds
|
| 41 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m INFO 02-03 01:42:00 [backends.py:805] Using cache directory: /home/mingzhe/.cache/vllm/torch_compile_cache/9c9322c6b2/rank_0_0/backbone for vLLM's torch.compile
|
| 42 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m INFO 02-03 01:42:00 [backends.py:865] Dynamo bytecode transform time: 16.15 s
|
| 43 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m INFO 02-03 01:42:20 [backends.py:302] Cache the graph of compile range (1, 8192) for later use
|
| 44 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m INFO 02-03 01:44:12 [backends.py:319] Compiling a graph for compile range (1, 8192) takes 120.43 s
|
| 45 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m INFO 02-03 01:44:12 [monitor.py:34] torch.compile takes 136.58 s in total
|
| 46 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] EngineCore failed to start.
|
| 47 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] Traceback (most recent call last):
|
| 48 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/worker/gpu_model_runner.py", line 4800, in _dummy_sampler_run
|
| 49 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] sampler_output = self.sampler(
|
| 50 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^
|
| 51 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
|
| 52 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] return self._call_impl(*args, **kwargs)
|
| 53 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 54 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
|
| 55 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] return forward_call(*args, **kwargs)
|
| 56 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 57 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/sample/sampler.py", line 96, in forward
|
| 58 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] sampled, processed_logprobs = self.sample(logits, sampling_metadata)
|
| 59 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 60 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/sample/sampler.py", line 187, in sample
|
| 61 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] random_sampled, processed_logprobs = self.topk_topp_sampler(
|
| 62 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^
|
| 63 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
|
| 64 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] return self._call_impl(*args, **kwargs)
|
| 65 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 66 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
|
| 67 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] return forward_call(*args, **kwargs)
|
| 68 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 69 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/sample/ops/topk_topp_sampler.py", line 104, in forward_native
|
| 70 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] logits = self.apply_top_k_top_p(logits, k, p)
|
| 71 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 72 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/sample/ops/topk_topp_sampler.py", line 262, in apply_top_k_top_p
|
| 73 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] logits_sort, logits_idx = logits.sort(dim=-1, descending=False)
|
| 74 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 75 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 1.74 GiB. GPU 0 has a total capacity of 139.80 GiB of which 368.44 MiB is free. Process 603285 has 4.61 GiB memory in use. Process 3310619 has 68.65 GiB memory in use. Including non-PyTorch memory, this process has 66.16 GiB memory in use. Of the allocated memory 65.02 GiB is allocated by PyTorch, and 416.55 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
|
| 76 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946]
|
| 77 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] The above exception was the direct cause of the following exception:
|
| 78 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946]
|
| 79 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] Traceback (most recent call last):
|
| 80 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 937, in run_engine_core
|
| 81 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] engine_core = EngineCoreProc(*args, engine_index=dp_rank, **kwargs)
|
| 82 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 83 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 691, in __init__
|
| 84 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] super().__init__(
|
| 85 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 112, in __init__
|
| 86 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] num_gpu_blocks, num_cpu_blocks, kv_cache_config = self._initialize_kv_caches(
|
| 87 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 88 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 242, in _initialize_kv_caches
|
| 89 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] available_gpu_memory = self.model_executor.determine_available_memory()
|
| 90 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 91 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/executor/abstract.py", line 126, in determine_available_memory
|
| 92 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] return self.collective_rpc("determine_available_memory")
|
| 93 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 94 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/executor/uniproc_executor.py", line 75, in collective_rpc
|
| 95 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] result = run_method(self.driver_worker, method, args, kwargs)
|
| 96 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 97 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/serial_utils.py", line 461, in run_method
|
| 98 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] return func(*args, **kwargs)
|
| 99 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^
|
| 100 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/torch/utils/_contextlib.py", line 120, in decorate_context
|
| 101 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] return func(*args, **kwargs)
|
| 102 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^
|
| 103 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/worker/gpu_worker.py", line 322, in determine_available_memory
|
| 104 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] self.model_runner.profile_run()
|
| 105 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/worker/gpu_model_runner.py", line 4981, in profile_run
|
| 106 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] output = self._dummy_sampler_run(last_hidden_states)
|
| 107 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 108 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/torch/utils/_contextlib.py", line 120, in decorate_context
|
| 109 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] return func(*args, **kwargs)
|
| 110 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^
|
| 111 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/worker/gpu_model_runner.py", line 4805, in _dummy_sampler_run
|
| 112 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] raise RuntimeError(
|
| 113 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ERROR 02-03 01:44:12 [core.py:946] RuntimeError: CUDA out of memory occurred when warming up sampler with 1024 dummy requests. Please try lowering `max_num_seqs` or `gpu_memory_utilization` when initializing the engine.
|
| 114 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m Process EngineCore_DP0:
|
| 115 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m Traceback (most recent call last):
|
| 116 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/worker/gpu_model_runner.py", line 4800, in _dummy_sampler_run
|
| 117 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m sampler_output = self.sampler(
|
| 118 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^
|
| 119 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
|
| 120 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m return self._call_impl(*args, **kwargs)
|
| 121 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 122 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
|
| 123 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m return forward_call(*args, **kwargs)
|
| 124 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 125 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/sample/sampler.py", line 96, in forward
|
| 126 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m sampled, processed_logprobs = self.sample(logits, sampling_metadata)
|
| 127 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 128 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/sample/sampler.py", line 187, in sample
|
| 129 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m random_sampled, processed_logprobs = self.topk_topp_sampler(
|
| 130 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^
|
| 131 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
|
| 132 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m return self._call_impl(*args, **kwargs)
|
| 133 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 134 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
|
| 135 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m return forward_call(*args, **kwargs)
|
| 136 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 137 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/sample/ops/topk_topp_sampler.py", line 104, in forward_native
|
| 138 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m logits = self.apply_top_k_top_p(logits, k, p)
|
| 139 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 140 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/sample/ops/topk_topp_sampler.py", line 262, in apply_top_k_top_p
|
| 141 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m logits_sort, logits_idx = logits.sort(dim=-1, descending=False)
|
| 142 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 143 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 1.74 GiB. GPU 0 has a total capacity of 139.80 GiB of which 368.44 MiB is free. Process 603285 has 4.61 GiB memory in use. Process 3310619 has 68.65 GiB memory in use. Including non-PyTorch memory, this process has 66.16 GiB memory in use. Of the allocated memory 65.02 GiB is allocated by PyTorch, and 416.55 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
|
| 144 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 145 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m The above exception was the direct cause of the following exception:
|
| 146 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m
|
| 147 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m Traceback (most recent call last):
|
| 148 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/usr/lib/python3.12/multiprocessing/process.py", line 314, in _bootstrap
|
| 149 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m self.run()
|
| 150 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/usr/lib/python3.12/multiprocessing/process.py", line 108, in run
|
| 151 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m self._target(*self._args, **self._kwargs)
|
| 152 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 950, in run_engine_core
|
| 153 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m raise e
|
| 154 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 937, in run_engine_core
|
| 155 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m engine_core = EngineCoreProc(*args, engine_index=dp_rank, **kwargs)
|
| 156 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 157 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 691, in __init__
|
| 158 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m super().__init__(
|
| 159 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 112, in __init__
|
| 160 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m num_gpu_blocks, num_cpu_blocks, kv_cache_config = self._initialize_kv_caches(
|
| 161 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 162 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 242, in _initialize_kv_caches
|
| 163 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m available_gpu_memory = self.model_executor.determine_available_memory()
|
| 164 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 165 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/executor/abstract.py", line 126, in determine_available_memory
|
| 166 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m return self.collective_rpc("determine_available_memory")
|
| 167 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 168 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/executor/uniproc_executor.py", line 75, in collective_rpc
|
| 169 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m result = run_method(self.driver_worker, method, args, kwargs)
|
| 170 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 171 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/serial_utils.py", line 461, in run_method
|
| 172 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m return func(*args, **kwargs)
|
| 173 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^
|
| 174 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/torch/utils/_contextlib.py", line 120, in decorate_context
|
| 175 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m return func(*args, **kwargs)
|
| 176 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^
|
| 177 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/worker/gpu_worker.py", line 322, in determine_available_memory
|
| 178 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m self.model_runner.profile_run()
|
| 179 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/worker/gpu_model_runner.py", line 4981, in profile_run
|
| 180 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m output = self._dummy_sampler_run(last_hidden_states)
|
| 181 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 182 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/torch/utils/_contextlib.py", line 120, in decorate_context
|
| 183 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m return func(*args, **kwargs)
|
| 184 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m ^^^^^^^^^^^^^^^^^^^^^
|
| 185 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/worker/gpu_model_runner.py", line 4805, in _dummy_sampler_run
|
| 186 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m raise RuntimeError(
|
| 187 |
+
[0;36m(EngineCore_DP0 pid=3312712)[0;0m RuntimeError: CUDA out of memory occurred when warming up sampler with 1024 dummy requests. Please try lowering `max_num_seqs` or `gpu_memory_utilization` when initializing the engine.
|
| 188 |
+
[rank0]:[W203 01:44:15.196915434 ProcessGroupNCCL.cpp:1524] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
|
| 189 |
+
[0;36m(APIServer pid=3309898)[0;0m Traceback (most recent call last):
|
| 190 |
+
[0;36m(APIServer pid=3309898)[0;0m File "<frozen runpy>", line 198, in _run_module_as_main
|
| 191 |
+
[0;36m(APIServer pid=3309898)[0;0m File "<frozen runpy>", line 88, in _run_code
|
| 192 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 991, in <module>
|
| 193 |
+
[0;36m(APIServer pid=3309898)[0;0m uvloop.run(run_server(args))
|
| 194 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/uvloop/__init__.py", line 96, in run
|
| 195 |
+
[0;36m(APIServer pid=3309898)[0;0m return __asyncio.run(
|
| 196 |
+
[0;36m(APIServer pid=3309898)[0;0m ^^^^^^^^^^^^^^
|
| 197 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/usr/lib/python3.12/asyncio/runners.py", line 194, in run
|
| 198 |
+
[0;36m(APIServer pid=3309898)[0;0m return runner.run(main)
|
| 199 |
+
[0;36m(APIServer pid=3309898)[0;0m ^^^^^^^^^^^^^^^^
|
| 200 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/usr/lib/python3.12/asyncio/runners.py", line 118, in run
|
| 201 |
+
[0;36m(APIServer pid=3309898)[0;0m return self._loop.run_until_complete(task)
|
| 202 |
+
[0;36m(APIServer pid=3309898)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 203 |
+
[0;36m(APIServer pid=3309898)[0;0m File "uvloop/loop.pyx", line 1518, in uvloop.loop.Loop.run_until_complete
|
| 204 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/uvloop/__init__.py", line 48, in wrapper
|
| 205 |
+
[0;36m(APIServer pid=3309898)[0;0m return await main
|
| 206 |
+
[0;36m(APIServer pid=3309898)[0;0m ^^^^^^^^^^
|
| 207 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 919, in run_server
|
| 208 |
+
[0;36m(APIServer pid=3309898)[0;0m await run_server_worker(listen_address, sock, args, **uvicorn_kwargs)
|
| 209 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 938, in run_server_worker
|
| 210 |
+
[0;36m(APIServer pid=3309898)[0;0m async with build_async_engine_client(
|
| 211 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/usr/lib/python3.12/contextlib.py", line 210, in __aenter__
|
| 212 |
+
[0;36m(APIServer pid=3309898)[0;0m return await anext(self.gen)
|
| 213 |
+
[0;36m(APIServer pid=3309898)[0;0m ^^^^^^^^^^^^^^^^^^^^^
|
| 214 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 147, in build_async_engine_client
|
| 215 |
+
[0;36m(APIServer pid=3309898)[0;0m async with build_async_engine_client_from_engine_args(
|
| 216 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/usr/lib/python3.12/contextlib.py", line 210, in __aenter__
|
| 217 |
+
[0;36m(APIServer pid=3309898)[0;0m return await anext(self.gen)
|
| 218 |
+
[0;36m(APIServer pid=3309898)[0;0m ^^^^^^^^^^^^^^^^^^^^^
|
| 219 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 188, in build_async_engine_client_from_engine_args
|
| 220 |
+
[0;36m(APIServer pid=3309898)[0;0m async_llm = AsyncLLM.from_vllm_config(
|
| 221 |
+
[0;36m(APIServer pid=3309898)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 222 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/async_llm.py", line 228, in from_vllm_config
|
| 223 |
+
[0;36m(APIServer pid=3309898)[0;0m return cls(
|
| 224 |
+
[0;36m(APIServer pid=3309898)[0;0m ^^^^
|
| 225 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/async_llm.py", line 155, in __init__
|
| 226 |
+
[0;36m(APIServer pid=3309898)[0;0m self.engine_core = EngineCoreClient.make_async_mp_client(
|
| 227 |
+
[0;36m(APIServer pid=3309898)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 228 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core_client.py", line 122, in make_async_mp_client
|
| 229 |
+
[0;36m(APIServer pid=3309898)[0;0m return AsyncMPClient(*client_args)
|
| 230 |
+
[0;36m(APIServer pid=3309898)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 231 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core_client.py", line 819, in __init__
|
| 232 |
+
[0;36m(APIServer pid=3309898)[0;0m super().__init__(
|
| 233 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core_client.py", line 479, in __init__
|
| 234 |
+
[0;36m(APIServer pid=3309898)[0;0m with launch_core_engines(vllm_config, executor_class, log_stats) as (
|
| 235 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/usr/lib/python3.12/contextlib.py", line 144, in __exit__
|
| 236 |
+
[0;36m(APIServer pid=3309898)[0;0m next(self.gen)
|
| 237 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/utils.py", line 933, in launch_core_engines
|
| 238 |
+
[0;36m(APIServer pid=3309898)[0;0m wait_for_engine_startup(
|
| 239 |
+
[0;36m(APIServer pid=3309898)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/utils.py", line 992, in wait_for_engine_startup
|
| 240 |
+
[0;36m(APIServer pid=3309898)[0;0m raise RuntimeError(
|
| 241 |
+
[0;36m(APIServer pid=3309898)[0;0m RuntimeError: Engine core initialization failed. See root cause above. Failed core proc(s): {}
|
vllm_0006000.log
CHANGED
|
@@ -1,20 +1,18 @@
|
|
| 1 |
-
[0;36m(APIServer pid=
|
| 2 |
-
[0;36m(APIServer pid=
|
| 3 |
-
[0;36m(APIServer pid=
|
| 4 |
-
[0;36m(APIServer pid=
|
| 5 |
-
[0;36m(APIServer pid=
|
| 6 |
-
[0;36m(APIServer pid=
|
| 7 |
-
[0;36m(APIServer pid=
|
| 8 |
-
[0;36m(APIServer pid=
|
| 9 |
-
[0;36m(APIServer pid=
|
| 10 |
-
[0;36m(APIServer pid=
|
| 11 |
-
[0;36m(APIServer pid=
|
| 12 |
-
[0;36m(APIServer pid=
|
| 13 |
-
[0;36m(APIServer pid=
|
| 14 |
-
[0;36m(EngineCore_DP0 pid=
|
| 15 |
-
[0;36m(EngineCore_DP0 pid=
|
| 16 |
-
[0;36m(EngineCore_DP0 pid=
|
| 17 |
-
[0;36m(EngineCore_DP0 pid=
|
| 18 |
-
[0;36m(EngineCore_DP0 pid=
|
| 19 |
-
[0;36m(EngineCore_DP0 pid=3235235)[0;0m INFO 02-03 01:27:14 [bitsandbytes_loader.py:786] Loading weights with BitsAndBytes quantization. May take a while ...
|
| 20 |
-
Cancellation requested; stopping current tasks.
|
|
|
|
| 1 |
+
[0;36m(APIServer pid=3310473)[0;0m INFO 02-03 01:41:05 [utils.py:325]
|
| 2 |
+
[0;36m(APIServer pid=3310473)[0;0m INFO 02-03 01:41:05 [utils.py:325] β β ββ ββ
|
| 3 |
+
[0;36m(APIServer pid=3310473)[0;0m INFO 02-03 01:41:05 [utils.py:325] ββ ββ β β β βββ β version 0.15.0
|
| 4 |
+
[0;36m(APIServer pid=3310473)[0;0m INFO 02-03 01:41:05 [utils.py:325] ββββ β β β β model Elfsong/VLM_stage_2_iter_0006000
|
| 5 |
+
[0;36m(APIServer pid=3310473)[0;0m INFO 02-03 01:41:05 [utils.py:325] ββ βββββ βββββ β β
|
| 6 |
+
[0;36m(APIServer pid=3310473)[0;0m INFO 02-03 01:41:05 [utils.py:325]
|
| 7 |
+
[0;36m(APIServer pid=3310473)[0;0m INFO 02-03 01:41:05 [utils.py:261] non-default args: {'port': 9002, 'model': 'Elfsong/VLM_stage_2_iter_0006000', 'trust_remote_code': True, 'gpu_memory_utilization': 0.4}
|
| 8 |
+
[0;36m(APIServer pid=3310473)[0;0m The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored.
|
| 9 |
+
[0;36m(APIServer pid=3310473)[0;0m The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored.
|
| 10 |
+
[0;36m(APIServer pid=3310473)[0;0m INFO 02-03 01:41:06 [model.py:541] Resolved architecture: Qwen3ForCausalLM
|
| 11 |
+
[0;36m(APIServer pid=3310473)[0;0m INFO 02-03 01:41:06 [model.py:1561] Using max model len 40960
|
| 12 |
+
[0;36m(APIServer pid=3310473)[0;0m INFO 02-03 01:41:06 [scheduler.py:226] Chunked prefill is enabled with max_num_batched_tokens=8192.
|
| 13 |
+
[0;36m(APIServer pid=3310473)[0;0m INFO 02-03 01:41:06 [vllm.py:624] Asynchronous scheduling is enabled.
|
| 14 |
+
[0;36m(EngineCore_DP0 pid=3313628)[0;0m INFO 02-03 01:41:18 [core.py:96] Initializing a V1 LLM engine (v0.15.0) with config: model='Elfsong/VLM_stage_2_iter_0006000', speculative_config=None, tokenizer='Elfsong/VLM_stage_2_iter_0006000', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=40960, download_dir=None, load_format=auto, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, enable_return_routed_experts=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser='', reasoning_parser_plugin='', enable_in_reasoning=False), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None, kv_cache_metrics=False, kv_cache_metrics_sample=0.01, cudagraph_metrics=False, enable_layerwise_nvtx_tracing=False, enable_mfu_metrics=False, enable_mm_processor_stats=False, enable_logging_iteration_details=False), seed=0, served_model_name=Elfsong/VLM_stage_2_iter_0006000, enable_prefix_caching=True, enable_chunked_prefill=True, pooler_config=None, compilation_config={'level': None, 'mode': <CompilationMode.VLLM_COMPILE: 3>, 'debug_dump_path': None, 'cache_dir': '', 'compile_cache_save_format': 'binary', 'backend': 'inductor', 'custom_ops': ['none'], 'splitting_ops': ['vllm::unified_attention', 'vllm::unified_attention_with_output', 'vllm::unified_mla_attention', 'vllm::unified_mla_attention_with_output', 'vllm::mamba_mixer2', 'vllm::mamba_mixer', 'vllm::short_conv', 'vllm::linear_attention', 'vllm::plamo2_mamba_mixer', 'vllm::gdn_attention_core', 'vllm::kda_attention', 'vllm::sparse_attn_indexer', 'vllm::rocm_aiter_sparse_attn_indexer'], 'compile_mm_encoder': False, 'compile_sizes': [], 'compile_ranges_split_points': [8192], 'inductor_compile_config': {'enable_auto_functionalized_v2': False, 'combo_kernels': True, 'benchmark_combo_kernel': True}, 'inductor_passes': {}, 'cudagraph_mode': <CUDAGraphMode.FULL_AND_PIECEWISE: (2, 1)>, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [1, 2, 4, 8, 16, 24, 32, 40, 48, 56, 64, 72, 80, 88, 96, 104, 112, 120, 128, 136, 144, 152, 160, 168, 176, 184, 192, 200, 208, 216, 224, 232, 240, 248, 256, 272, 288, 304, 320, 336, 352, 368, 384, 400, 416, 432, 448, 464, 480, 496, 512], 'cudagraph_copy_inputs': False, 'cudagraph_specialize_lora': True, 'use_inductor_graph_partition': False, 'pass_config': {'fuse_norm_quant': False, 'fuse_act_quant': False, 'fuse_attn_quant': False, 'eliminate_noops': True, 'enable_sp': False, 'fuse_gemm_comms': False, 'fuse_allreduce_rms': False}, 'max_cudagraph_capture_size': 512, 'dynamic_shapes_config': {'type': <DynamicShapesType.BACKED: 'backed'>, 'evaluate_guards': False, 'assume_32_bit_indexing': True}, 'local_cache_dir': None}
|
| 15 |
+
[0;36m(EngineCore_DP0 pid=3313628)[0;0m INFO 02-03 01:41:21 [parallel_state.py:1212] world_size=1 rank=0 local_rank=0 distributed_init_method=tcp://172.21.25.98:47213 backend=nccl
|
| 16 |
+
[0;36m(EngineCore_DP0 pid=3313628)[0;0m INFO 02-03 01:41:21 [parallel_state.py:1423] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, PCP rank 0, TP rank 0, EP rank N/A
|
| 17 |
+
[0;36m(EngineCore_DP0 pid=3313628)[0;0m INFO 02-03 01:41:23 [gpu_model_runner.py:4021] Starting to load model Elfsong/VLM_stage_2_iter_0006000...
|
| 18 |
+
[0;36m(EngineCore_DP0 pid=3313628)[0;0m INFO 02-03 01:41:24 [cuda.py:364] Using FLASH_ATTN attention backend out of potential backends: ('FLASH_ATTN', 'FLASHINFER', 'TRITON_ATTN', 'FLEX_ATTENTION')
|
|
|
|
|
|
vllm_0006500.log
CHANGED
|
@@ -1,97 +1,203 @@
|
|
| 1 |
-
[0;36m(APIServer pid=
|
| 2 |
-
[0;36m(APIServer pid=
|
| 3 |
-
[0;36m(APIServer pid=
|
| 4 |
-
[0;36m(APIServer pid=
|
| 5 |
-
[0;36m(APIServer pid=
|
| 6 |
-
[0;36m(APIServer pid=
|
| 7 |
-
[0;36m(
|
| 8 |
-
[0;36m(
|
| 9 |
-
[0;36m(
|
| 10 |
-
[0;36m(
|
| 11 |
-
[0;36m(
|
| 12 |
-
[0;36m(
|
| 13 |
-
[0;36m(
|
| 14 |
-
[0;36m(EngineCore_DP0 pid=
|
| 15 |
-
[0;36m(EngineCore_DP0 pid=
|
| 16 |
-
[0;36m(EngineCore_DP0 pid=
|
| 17 |
-
[0;36m(EngineCore_DP0 pid=
|
| 18 |
-
[0;36m(EngineCore_DP0 pid=
|
| 19 |
-
[0;36m(EngineCore_DP0 pid=
|
| 20 |
-
[0;36m(EngineCore_DP0 pid=
|
| 21 |
-
[0;36m(EngineCore_DP0 pid=
|
| 22 |
-
[0;36m(EngineCore_DP0 pid=
|
| 23 |
-
[0;36m(EngineCore_DP0 pid=
|
| 24 |
-
[0;36m(EngineCore_DP0 pid=
|
| 25 |
-
[0;36m(EngineCore_DP0 pid=
|
| 26 |
-
[0;36m(EngineCore_DP0 pid=
|
| 27 |
-
[0;36m(EngineCore_DP0 pid=
|
| 28 |
-
[0;36m(EngineCore_DP0 pid=
|
| 29 |
-
[0;36m(EngineCore_DP0 pid=
|
| 30 |
-
[0;36m(EngineCore_DP0 pid=
|
| 31 |
-
[0;36m(EngineCore_DP0 pid=
|
| 32 |
-
[0;36m(EngineCore_DP0 pid=
|
| 33 |
-
[0;36m(EngineCore_DP0 pid=
|
| 34 |
-
[0;36m(EngineCore_DP0 pid=
|
| 35 |
-
[0;36m(EngineCore_DP0 pid=
|
| 36 |
-
[0;36m(EngineCore_DP0 pid=
|
| 37 |
-
[0;36m(EngineCore_DP0 pid=
|
| 38 |
-
[0;36m(EngineCore_DP0 pid=
|
| 39 |
-
[0;36m(EngineCore_DP0 pid=
|
| 40 |
-
[0;36m(EngineCore_DP0 pid=
|
| 41 |
-
[0;36m(EngineCore_DP0 pid=
|
| 42 |
-
[0;36m(EngineCore_DP0 pid=
|
| 43 |
-
[0;36m(EngineCore_DP0 pid=
|
| 44 |
-
[0;36m(EngineCore_DP0 pid=
|
| 45 |
-
[0;36m(EngineCore_DP0 pid=
|
| 46 |
-
[0;36m(
|
| 47 |
-
[0;36m(
|
| 48 |
-
[0;36m(
|
| 49 |
-
[0;36m(
|
| 50 |
-
[0;36m(
|
| 51 |
-
[0;36m(
|
| 52 |
-
[0;36m(
|
| 53 |
-
[0;36m(
|
| 54 |
-
[0;36m(
|
| 55 |
-
[0;36m(
|
| 56 |
-
[0;36m(
|
| 57 |
-
[0;36m(
|
| 58 |
-
[0;36m(
|
| 59 |
-
[0;36m(
|
| 60 |
-
[0;36m(
|
| 61 |
-
[0;36m(
|
| 62 |
-
[0;36m(
|
| 63 |
-
[0;36m(
|
| 64 |
-
[0;36m(
|
| 65 |
-
[0;36m(
|
| 66 |
-
[0;36m(
|
| 67 |
-
[0;36m(
|
| 68 |
-
[0;36m(
|
| 69 |
-
[0;36m(
|
| 70 |
-
[0;36m(
|
| 71 |
-
[0;36m(
|
| 72 |
-
[0;36m(
|
| 73 |
-
[0;36m(
|
| 74 |
-
[0;36m(
|
| 75 |
-
[0;36m(
|
| 76 |
-
[0;36m(
|
| 77 |
-
[0;36m(
|
| 78 |
-
[0;36m(
|
| 79 |
-
[0;36m(
|
| 80 |
-
[0;36m(
|
| 81 |
-
[0;36m(
|
| 82 |
-
[0;36m(
|
| 83 |
-
[0;36m(
|
| 84 |
-
[0;36m(
|
| 85 |
-
[0;36m(
|
| 86 |
-
[0;36m(
|
| 87 |
-
[0;36m(
|
| 88 |
-
[0;36m(
|
| 89 |
-
[0;36m(
|
| 90 |
-
[0;36m(
|
| 91 |
-
[0;36m(
|
| 92 |
-
[0;36m(
|
| 93 |
-
[0;36m(
|
| 94 |
-
[
|
| 95 |
-
[0;36m(
|
| 96 |
-
[0;36m(
|
| 97 |
-
[0;36m(
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[0;36m(APIServer pid=3311584)[0;0m INFO 02-03 01:41:10 [utils.py:325]
|
| 2 |
+
[0;36m(APIServer pid=3311584)[0;0m INFO 02-03 01:41:10 [utils.py:325] β β ββ ββ
|
| 3 |
+
[0;36m(APIServer pid=3311584)[0;0m INFO 02-03 01:41:10 [utils.py:325] ββ ββ β β β βββ β version 0.15.0
|
| 4 |
+
[0;36m(APIServer pid=3311584)[0;0m INFO 02-03 01:41:10 [utils.py:325] ββββ β β β β model Elfsong/VLM_stage_2_iter_0006500
|
| 5 |
+
[0;36m(APIServer pid=3311584)[0;0m INFO 02-03 01:41:10 [utils.py:325] ββ βββββ βββββ β β
|
| 6 |
+
[0;36m(APIServer pid=3311584)[0;0m INFO 02-03 01:41:10 [utils.py:325]
|
| 7 |
+
[0;36m(APIServer pid=3311584)[0;0m INFO 02-03 01:41:10 [utils.py:261] non-default args: {'port': 9003, 'model': 'Elfsong/VLM_stage_2_iter_0006500', 'trust_remote_code': True, 'gpu_memory_utilization': 0.4}
|
| 8 |
+
[0;36m(APIServer pid=3311584)[0;0m The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored.
|
| 9 |
+
[0;36m(APIServer pid=3311584)[0;0m The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored.
|
| 10 |
+
[0;36m(APIServer pid=3311584)[0;0m INFO 02-03 01:41:12 [model.py:541] Resolved architecture: Qwen3ForCausalLM
|
| 11 |
+
[0;36m(APIServer pid=3311584)[0;0m INFO 02-03 01:41:12 [model.py:1561] Using max model len 40960
|
| 12 |
+
[0;36m(APIServer pid=3311584)[0;0m INFO 02-03 01:41:12 [scheduler.py:226] Chunked prefill is enabled with max_num_batched_tokens=8192.
|
| 13 |
+
[0;36m(APIServer pid=3311584)[0;0m INFO 02-03 01:41:12 [vllm.py:624] Asynchronous scheduling is enabled.
|
| 14 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m INFO 02-03 01:41:31 [core.py:96] Initializing a V1 LLM engine (v0.15.0) with config: model='Elfsong/VLM_stage_2_iter_0006500', speculative_config=None, tokenizer='Elfsong/VLM_stage_2_iter_0006500', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=40960, download_dir=None, load_format=auto, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, enable_return_routed_experts=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser='', reasoning_parser_plugin='', enable_in_reasoning=False), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None, kv_cache_metrics=False, kv_cache_metrics_sample=0.01, cudagraph_metrics=False, enable_layerwise_nvtx_tracing=False, enable_mfu_metrics=False, enable_mm_processor_stats=False, enable_logging_iteration_details=False), seed=0, served_model_name=Elfsong/VLM_stage_2_iter_0006500, enable_prefix_caching=True, enable_chunked_prefill=True, pooler_config=None, compilation_config={'level': None, 'mode': <CompilationMode.VLLM_COMPILE: 3>, 'debug_dump_path': None, 'cache_dir': '', 'compile_cache_save_format': 'binary', 'backend': 'inductor', 'custom_ops': ['none'], 'splitting_ops': ['vllm::unified_attention', 'vllm::unified_attention_with_output', 'vllm::unified_mla_attention', 'vllm::unified_mla_attention_with_output', 'vllm::mamba_mixer2', 'vllm::mamba_mixer', 'vllm::short_conv', 'vllm::linear_attention', 'vllm::plamo2_mamba_mixer', 'vllm::gdn_attention_core', 'vllm::kda_attention', 'vllm::sparse_attn_indexer', 'vllm::rocm_aiter_sparse_attn_indexer'], 'compile_mm_encoder': False, 'compile_sizes': [], 'compile_ranges_split_points': [8192], 'inductor_compile_config': {'enable_auto_functionalized_v2': False, 'combo_kernels': True, 'benchmark_combo_kernel': True}, 'inductor_passes': {}, 'cudagraph_mode': <CUDAGraphMode.FULL_AND_PIECEWISE: (2, 1)>, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [1, 2, 4, 8, 16, 24, 32, 40, 48, 56, 64, 72, 80, 88, 96, 104, 112, 120, 128, 136, 144, 152, 160, 168, 176, 184, 192, 200, 208, 216, 224, 232, 240, 248, 256, 272, 288, 304, 320, 336, 352, 368, 384, 400, 416, 432, 448, 464, 480, 496, 512], 'cudagraph_copy_inputs': False, 'cudagraph_specialize_lora': True, 'use_inductor_graph_partition': False, 'pass_config': {'fuse_norm_quant': False, 'fuse_act_quant': False, 'fuse_attn_quant': False, 'eliminate_noops': True, 'enable_sp': False, 'fuse_gemm_comms': False, 'fuse_allreduce_rms': False}, 'max_cudagraph_capture_size': 512, 'dynamic_shapes_config': {'type': <DynamicShapesType.BACKED: 'backed'>, 'evaluate_guards': False, 'assume_32_bit_indexing': True}, 'local_cache_dir': None}
|
| 15 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m INFO 02-03 01:41:35 [parallel_state.py:1212] world_size=1 rank=0 local_rank=0 distributed_init_method=tcp://172.21.25.98:59489 backend=nccl
|
| 16 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m INFO 02-03 01:41:35 [parallel_state.py:1423] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, PCP rank 0, TP rank 0, EP rank N/A
|
| 17 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m INFO 02-03 01:41:37 [gpu_model_runner.py:4021] Starting to load model Elfsong/VLM_stage_2_iter_0006500...
|
| 18 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m INFO 02-03 01:41:38 [cuda.py:364] Using FLASH_ATTN attention backend out of potential backends: ('FLASH_ATTN', 'FLASHINFER', 'TRITON_ATTN', 'FLEX_ATTENTION')
|
| 19 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [gpu_model_runner.py:4116] Failed to load model - not enough GPU memory. Try lowering --gpu-memory-utilization to free memory for weights, increasing --tensor-parallel-size, or using --quantization. See https://docs.vllm.ai/en/latest/configuration/conserving_memory/ for more tips. (original error: CUDA out of memory. Tried to allocate 500.00 MiB. GPU 0 has a total capacity of 139.80 GiB of which 312.25 MiB is free. Process 3284142 has 5.32 GiB memory in use. Process 3243122 has 2.91 GiB memory in use. Process 3243120 has 2.80 GiB memory in use. Process 3243121 has 3.62 GiB memory in use. Process 3243119 has 3.37 GiB memory in use. Process 3315388 has 61.82 GiB memory in use. Including non-PyTorch memory, this process has 59.60 GiB memory in use. Of the allocated memory 58.94 GiB is allocated by PyTorch, and 278.00 KiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables))
|
| 20 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] EngineCore failed to start.
|
| 21 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] Traceback (most recent call last):
|
| 22 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 937, in run_engine_core
|
| 23 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] engine_core = EngineCoreProc(*args, engine_index=dp_rank, **kwargs)
|
| 24 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 25 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 691, in __init__
|
| 26 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] super().__init__(
|
| 27 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 105, in __init__
|
| 28 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] self.model_executor = executor_class(vllm_config)
|
| 29 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 30 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/executor/abstract.py", line 101, in __init__
|
| 31 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] self._init_executor()
|
| 32 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/executor/uniproc_executor.py", line 48, in _init_executor
|
| 33 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] self.driver_worker.load_model()
|
| 34 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/worker/gpu_worker.py", line 275, in load_model
|
| 35 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] self.model_runner.load_model(eep_scale_up=eep_scale_up)
|
| 36 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/worker/gpu_model_runner.py", line 4117, in load_model
|
| 37 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] raise e
|
| 38 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/worker/gpu_model_runner.py", line 4040, in load_model
|
| 39 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] self.model = model_loader.load_model(
|
| 40 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^
|
| 41 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/model_loader/base_loader.py", line 50, in load_model
|
| 42 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] model = initialize_model(
|
| 43 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] ^^^^^^^^^^^^^^^^^
|
| 44 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/model_loader/utils.py", line 48, in initialize_model
|
| 45 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] return model_class(vllm_config=vllm_config, prefix=prefix)
|
| 46 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 47 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/models/qwen3.py", line 274, in __init__
|
| 48 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] self.model = Qwen3Model(
|
| 49 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] ^^^^^^^^^^^
|
| 50 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/compilation/decorators.py", line 306, in __init__
|
| 51 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] old_init(self, **kwargs)
|
| 52 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/models/qwen3.py", line 248, in __init__
|
| 53 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] super().__init__(
|
| 54 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/compilation/decorators.py", line 306, in __init__
|
| 55 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] old_init(self, **kwargs)
|
| 56 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/models/qwen2.py", line 394, in __init__
|
| 57 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] self.start_layer, self.end_layer, self.layers = make_layers(
|
| 58 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] ^^^^^^^^^^^^
|
| 59 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/models/utils.py", line 707, in make_layers
|
| 60 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] maybe_offload_to_cpu(layer_fn(prefix=f"{prefix}.{idx}"))
|
| 61 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 62 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/models/qwen2.py", line 396, in <lambda>
|
| 63 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] lambda prefix: decoder_layer_type(
|
| 64 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] ^^^^^^^^^^^^^^^^^^^
|
| 65 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/models/qwen3.py", line 196, in __init__
|
| 66 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] self.mlp = Qwen3MLP(
|
| 67 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] ^^^^^^^^^
|
| 68 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/models/qwen2.py", line 87, in __init__
|
| 69 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] self.gate_up_proj = MergedColumnParallelLinear(
|
| 70 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 71 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/layers/linear.py", line 670, in __init__
|
| 72 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] super().__init__(
|
| 73 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/layers/linear.py", line 495, in __init__
|
| 74 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] self.quant_method.create_weights(
|
| 75 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/layers/linear.py", line 224, in create_weights
|
| 76 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] data=torch.empty(
|
| 77 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] ^^^^^^^^^^^^
|
| 78 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/torch/utils/_device.py", line 103, in __torch_function__
|
| 79 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] return func(*args, **kwargs)
|
| 80 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] ^^^^^^^^^^^^^^^^^^^^^
|
| 81 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ERROR 02-03 01:41:41 [core.py:946] torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 500.00 MiB. GPU 0 has a total capacity of 139.80 GiB of which 312.25 MiB is free. Process 3284142 has 5.32 GiB memory in use. Process 3243122 has 2.91 GiB memory in use. Process 3243120 has 2.80 GiB memory in use. Process 3243121 has 3.62 GiB memory in use. Process 3243119 has 3.37 GiB memory in use. Process 3315388 has 61.82 GiB memory in use. Including non-PyTorch memory, this process has 59.60 GiB memory in use. Of the allocated memory 58.94 GiB is allocated by PyTorch, and 278.00 KiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
|
| 82 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m Process EngineCore_DP0:
|
| 83 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m Traceback (most recent call last):
|
| 84 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/usr/lib/python3.12/multiprocessing/process.py", line 314, in _bootstrap
|
| 85 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m self.run()
|
| 86 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/usr/lib/python3.12/multiprocessing/process.py", line 108, in run
|
| 87 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m self._target(*self._args, **self._kwargs)
|
| 88 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 950, in run_engine_core
|
| 89 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m raise e
|
| 90 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 937, in run_engine_core
|
| 91 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m engine_core = EngineCoreProc(*args, engine_index=dp_rank, **kwargs)
|
| 92 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 93 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 691, in __init__
|
| 94 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m super().__init__(
|
| 95 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 105, in __init__
|
| 96 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m self.model_executor = executor_class(vllm_config)
|
| 97 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 98 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/executor/abstract.py", line 101, in __init__
|
| 99 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m self._init_executor()
|
| 100 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/executor/uniproc_executor.py", line 48, in _init_executor
|
| 101 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m self.driver_worker.load_model()
|
| 102 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/worker/gpu_worker.py", line 275, in load_model
|
| 103 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m self.model_runner.load_model(eep_scale_up=eep_scale_up)
|
| 104 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/worker/gpu_model_runner.py", line 4117, in load_model
|
| 105 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m raise e
|
| 106 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/worker/gpu_model_runner.py", line 4040, in load_model
|
| 107 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m self.model = model_loader.load_model(
|
| 108 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^
|
| 109 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/model_loader/base_loader.py", line 50, in load_model
|
| 110 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m model = initialize_model(
|
| 111 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ^^^^^^^^^^^^^^^^^
|
| 112 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/model_loader/utils.py", line 48, in initialize_model
|
| 113 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m return model_class(vllm_config=vllm_config, prefix=prefix)
|
| 114 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 115 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/models/qwen3.py", line 274, in __init__
|
| 116 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m self.model = Qwen3Model(
|
| 117 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ^^^^^^^^^^^
|
| 118 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/compilation/decorators.py", line 306, in __init__
|
| 119 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m old_init(self, **kwargs)
|
| 120 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/models/qwen3.py", line 248, in __init__
|
| 121 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m super().__init__(
|
| 122 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/compilation/decorators.py", line 306, in __init__
|
| 123 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m old_init(self, **kwargs)
|
| 124 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/models/qwen2.py", line 394, in __init__
|
| 125 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m self.start_layer, self.end_layer, self.layers = make_layers(
|
| 126 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ^^^^^^^^^^^^
|
| 127 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/models/utils.py", line 707, in make_layers
|
| 128 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m maybe_offload_to_cpu(layer_fn(prefix=f"{prefix}.{idx}"))
|
| 129 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 130 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/models/qwen2.py", line 396, in <lambda>
|
| 131 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m lambda prefix: decoder_layer_type(
|
| 132 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ^^^^^^^^^^^^^^^^^^^
|
| 133 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/models/qwen3.py", line 196, in __init__
|
| 134 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m self.mlp = Qwen3MLP(
|
| 135 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ^^^^^^^^^
|
| 136 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/models/qwen2.py", line 87, in __init__
|
| 137 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m self.gate_up_proj = MergedColumnParallelLinear(
|
| 138 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 139 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/layers/linear.py", line 670, in __init__
|
| 140 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m super().__init__(
|
| 141 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/layers/linear.py", line 495, in __init__
|
| 142 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m self.quant_method.create_weights(
|
| 143 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/model_executor/layers/linear.py", line 224, in create_weights
|
| 144 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m data=torch.empty(
|
| 145 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ^^^^^^^^^^^^
|
| 146 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/torch/utils/_device.py", line 103, in __torch_function__
|
| 147 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m return func(*args, **kwargs)
|
| 148 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m ^^^^^^^^^^^^^^^^^^^^^
|
| 149 |
+
[0;36m(EngineCore_DP0 pid=3315747)[0;0m torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 500.00 MiB. GPU 0 has a total capacity of 139.80 GiB of which 312.25 MiB is free. Process 3284142 has 5.32 GiB memory in use. Process 3243122 has 2.91 GiB memory in use. Process 3243120 has 2.80 GiB memory in use. Process 3243121 has 3.62 GiB memory in use. Process 3243119 has 3.37 GiB memory in use. Process 3315388 has 61.82 GiB memory in use. Including non-PyTorch memory, this process has 59.60 GiB memory in use. Of the allocated memory 58.94 GiB is allocated by PyTorch, and 278.00 KiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
|
| 150 |
+
[rank0]:[W203 01:41:43.753415460 ProcessGroupNCCL.cpp:1524] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
|
| 151 |
+
[0;36m(APIServer pid=3311584)[0;0m Traceback (most recent call last):
|
| 152 |
+
[0;36m(APIServer pid=3311584)[0;0m File "<frozen runpy>", line 198, in _run_module_as_main
|
| 153 |
+
[0;36m(APIServer pid=3311584)[0;0m File "<frozen runpy>", line 88, in _run_code
|
| 154 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 991, in <module>
|
| 155 |
+
[0;36m(APIServer pid=3311584)[0;0m uvloop.run(run_server(args))
|
| 156 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/uvloop/__init__.py", line 96, in run
|
| 157 |
+
[0;36m(APIServer pid=3311584)[0;0m return __asyncio.run(
|
| 158 |
+
[0;36m(APIServer pid=3311584)[0;0m ^^^^^^^^^^^^^^
|
| 159 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/usr/lib/python3.12/asyncio/runners.py", line 194, in run
|
| 160 |
+
[0;36m(APIServer pid=3311584)[0;0m return runner.run(main)
|
| 161 |
+
[0;36m(APIServer pid=3311584)[0;0m ^^^^^^^^^^^^^^^^
|
| 162 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/usr/lib/python3.12/asyncio/runners.py", line 118, in run
|
| 163 |
+
[0;36m(APIServer pid=3311584)[0;0m return self._loop.run_until_complete(task)
|
| 164 |
+
[0;36m(APIServer pid=3311584)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 165 |
+
[0;36m(APIServer pid=3311584)[0;0m File "uvloop/loop.pyx", line 1518, in uvloop.loop.Loop.run_until_complete
|
| 166 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/uvloop/__init__.py", line 48, in wrapper
|
| 167 |
+
[0;36m(APIServer pid=3311584)[0;0m return await main
|
| 168 |
+
[0;36m(APIServer pid=3311584)[0;0m ^^^^^^^^^^
|
| 169 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 919, in run_server
|
| 170 |
+
[0;36m(APIServer pid=3311584)[0;0m await run_server_worker(listen_address, sock, args, **uvicorn_kwargs)
|
| 171 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 938, in run_server_worker
|
| 172 |
+
[0;36m(APIServer pid=3311584)[0;0m async with build_async_engine_client(
|
| 173 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/usr/lib/python3.12/contextlib.py", line 210, in __aenter__
|
| 174 |
+
[0;36m(APIServer pid=3311584)[0;0m return await anext(self.gen)
|
| 175 |
+
[0;36m(APIServer pid=3311584)[0;0m ^^^^^^^^^^^^^^^^^^^^^
|
| 176 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 147, in build_async_engine_client
|
| 177 |
+
[0;36m(APIServer pid=3311584)[0;0m async with build_async_engine_client_from_engine_args(
|
| 178 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/usr/lib/python3.12/contextlib.py", line 210, in __aenter__
|
| 179 |
+
[0;36m(APIServer pid=3311584)[0;0m return await anext(self.gen)
|
| 180 |
+
[0;36m(APIServer pid=3311584)[0;0m ^^^^^^^^^^^^^^^^^^^^^
|
| 181 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 188, in build_async_engine_client_from_engine_args
|
| 182 |
+
[0;36m(APIServer pid=3311584)[0;0m async_llm = AsyncLLM.from_vllm_config(
|
| 183 |
+
[0;36m(APIServer pid=3311584)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 184 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/async_llm.py", line 228, in from_vllm_config
|
| 185 |
+
[0;36m(APIServer pid=3311584)[0;0m return cls(
|
| 186 |
+
[0;36m(APIServer pid=3311584)[0;0m ^^^^
|
| 187 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/async_llm.py", line 155, in __init__
|
| 188 |
+
[0;36m(APIServer pid=3311584)[0;0m self.engine_core = EngineCoreClient.make_async_mp_client(
|
| 189 |
+
[0;36m(APIServer pid=3311584)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 190 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core_client.py", line 122, in make_async_mp_client
|
| 191 |
+
[0;36m(APIServer pid=3311584)[0;0m return AsyncMPClient(*client_args)
|
| 192 |
+
[0;36m(APIServer pid=3311584)[0;0m ^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 193 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core_client.py", line 819, in __init__
|
| 194 |
+
[0;36m(APIServer pid=3311584)[0;0m super().__init__(
|
| 195 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/core_client.py", line 479, in __init__
|
| 196 |
+
[0;36m(APIServer pid=3311584)[0;0m with launch_core_engines(vllm_config, executor_class, log_stats) as (
|
| 197 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/usr/lib/python3.12/contextlib.py", line 144, in __exit__
|
| 198 |
+
[0;36m(APIServer pid=3311584)[0;0m next(self.gen)
|
| 199 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/utils.py", line 933, in launch_core_engines
|
| 200 |
+
[0;36m(APIServer pid=3311584)[0;0m wait_for_engine_startup(
|
| 201 |
+
[0;36m(APIServer pid=3311584)[0;0m File "/home/mingzhe/Projects/Arena/.venv/lib/python3.12/site-packages/vllm/v1/engine/utils.py", line 992, in wait_for_engine_startup
|
| 202 |
+
[0;36m(APIServer pid=3311584)[0;0m raise RuntimeError(
|
| 203 |
+
[0;36m(APIServer pid=3311584)[0;0m RuntimeError: Engine core initialization failed. See root cause above. Failed core proc(s): {}
|
vllm_0007000.log
CHANGED
|
@@ -1,20 +1,18 @@
|
|
| 1 |
-
[0;36m(APIServer pid=
|
| 2 |
-
[0;36m(APIServer pid=
|
| 3 |
-
[0;36m(APIServer pid=
|
| 4 |
-
[0;36m(APIServer pid=
|
| 5 |
-
[0;36m(APIServer pid=
|
| 6 |
-
[0;36m(APIServer pid=
|
| 7 |
-
[0;36m(APIServer pid=
|
| 8 |
-
[0;36m(APIServer pid=
|
| 9 |
-
[0;36m(APIServer pid=
|
| 10 |
-
[0;36m(APIServer pid=
|
| 11 |
-
[0;36m(APIServer pid=
|
| 12 |
-
[0;36m(APIServer pid=
|
| 13 |
-
[0;36m(APIServer pid=
|
| 14 |
-
[0;36m(EngineCore_DP0 pid=
|
| 15 |
-
[0;36m(EngineCore_DP0 pid=
|
| 16 |
-
[0;36m(EngineCore_DP0 pid=
|
| 17 |
-
[0;36m(EngineCore_DP0 pid=
|
| 18 |
-
[0;36m(EngineCore_DP0 pid=
|
| 19 |
-
[0;36m(EngineCore_DP0 pid=3236204)[0;0m INFO 02-03 01:27:19 [bitsandbytes_loader.py:786] Loading weights with BitsAndBytes quantization. May take a while ...
|
| 20 |
-
Cancellation requested; stopping current tasks.
|
|
|
|
| 1 |
+
[0;36m(APIServer pid=3313448)[0;0m INFO 02-03 01:41:15 [utils.py:325]
|
| 2 |
+
[0;36m(APIServer pid=3313448)[0;0m INFO 02-03 01:41:15 [utils.py:325] β β ββ ββ
|
| 3 |
+
[0;36m(APIServer pid=3313448)[0;0m INFO 02-03 01:41:15 [utils.py:325] ββ ββ β β β βββ β version 0.15.0
|
| 4 |
+
[0;36m(APIServer pid=3313448)[0;0m INFO 02-03 01:41:15 [utils.py:325] ββββ β β β β model Elfsong/VLM_stage_2_iter_0007000
|
| 5 |
+
[0;36m(APIServer pid=3313448)[0;0m INFO 02-03 01:41:15 [utils.py:325] ββ βββββ βββββ β β
|
| 6 |
+
[0;36m(APIServer pid=3313448)[0;0m INFO 02-03 01:41:15 [utils.py:325]
|
| 7 |
+
[0;36m(APIServer pid=3313448)[0;0m INFO 02-03 01:41:15 [utils.py:261] non-default args: {'port': 9004, 'model': 'Elfsong/VLM_stage_2_iter_0007000', 'trust_remote_code': True, 'gpu_memory_utilization': 0.4}
|
| 8 |
+
[0;36m(APIServer pid=3313448)[0;0m The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored.
|
| 9 |
+
[0;36m(APIServer pid=3313448)[0;0m The argument `trust_remote_code` is to be used with Auto classes. It has no effect here and is ignored.
|
| 10 |
+
[0;36m(APIServer pid=3313448)[0;0m INFO 02-03 01:41:16 [model.py:541] Resolved architecture: Qwen3ForCausalLM
|
| 11 |
+
[0;36m(APIServer pid=3313448)[0;0m INFO 02-03 01:41:16 [model.py:1561] Using max model len 40960
|
| 12 |
+
[0;36m(APIServer pid=3313448)[0;0m INFO 02-03 01:41:16 [scheduler.py:226] Chunked prefill is enabled with max_num_batched_tokens=8192.
|
| 13 |
+
[0;36m(APIServer pid=3313448)[0;0m INFO 02-03 01:41:16 [vllm.py:624] Asynchronous scheduling is enabled.
|
| 14 |
+
[0;36m(EngineCore_DP0 pid=3315388)[0;0m INFO 02-03 01:41:28 [core.py:96] Initializing a V1 LLM engine (v0.15.0) with config: model='Elfsong/VLM_stage_2_iter_0007000', speculative_config=None, tokenizer='Elfsong/VLM_stage_2_iter_0007000', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=40960, download_dir=None, load_format=auto, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=False, enable_return_routed_experts=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser='', reasoning_parser_plugin='', enable_in_reasoning=False), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None, kv_cache_metrics=False, kv_cache_metrics_sample=0.01, cudagraph_metrics=False, enable_layerwise_nvtx_tracing=False, enable_mfu_metrics=False, enable_mm_processor_stats=False, enable_logging_iteration_details=False), seed=0, served_model_name=Elfsong/VLM_stage_2_iter_0007000, enable_prefix_caching=True, enable_chunked_prefill=True, pooler_config=None, compilation_config={'level': None, 'mode': <CompilationMode.VLLM_COMPILE: 3>, 'debug_dump_path': None, 'cache_dir': '', 'compile_cache_save_format': 'binary', 'backend': 'inductor', 'custom_ops': ['none'], 'splitting_ops': ['vllm::unified_attention', 'vllm::unified_attention_with_output', 'vllm::unified_mla_attention', 'vllm::unified_mla_attention_with_output', 'vllm::mamba_mixer2', 'vllm::mamba_mixer', 'vllm::short_conv', 'vllm::linear_attention', 'vllm::plamo2_mamba_mixer', 'vllm::gdn_attention_core', 'vllm::kda_attention', 'vllm::sparse_attn_indexer', 'vllm::rocm_aiter_sparse_attn_indexer'], 'compile_mm_encoder': False, 'compile_sizes': [], 'compile_ranges_split_points': [8192], 'inductor_compile_config': {'enable_auto_functionalized_v2': False, 'combo_kernels': True, 'benchmark_combo_kernel': True}, 'inductor_passes': {}, 'cudagraph_mode': <CUDAGraphMode.FULL_AND_PIECEWISE: (2, 1)>, 'cudagraph_num_of_warmups': 1, 'cudagraph_capture_sizes': [1, 2, 4, 8, 16, 24, 32, 40, 48, 56, 64, 72, 80, 88, 96, 104, 112, 120, 128, 136, 144, 152, 160, 168, 176, 184, 192, 200, 208, 216, 224, 232, 240, 248, 256, 272, 288, 304, 320, 336, 352, 368, 384, 400, 416, 432, 448, 464, 480, 496, 512], 'cudagraph_copy_inputs': False, 'cudagraph_specialize_lora': True, 'use_inductor_graph_partition': False, 'pass_config': {'fuse_norm_quant': False, 'fuse_act_quant': False, 'fuse_attn_quant': False, 'eliminate_noops': True, 'enable_sp': False, 'fuse_gemm_comms': False, 'fuse_allreduce_rms': False}, 'max_cudagraph_capture_size': 512, 'dynamic_shapes_config': {'type': <DynamicShapesType.BACKED: 'backed'>, 'evaluate_guards': False, 'assume_32_bit_indexing': True}, 'local_cache_dir': None}
|
| 15 |
+
[0;36m(EngineCore_DP0 pid=3315388)[0;0m INFO 02-03 01:41:32 [parallel_state.py:1212] world_size=1 rank=0 local_rank=0 distributed_init_method=tcp://172.21.25.98:53245 backend=nccl
|
| 16 |
+
[0;36m(EngineCore_DP0 pid=3315388)[0;0m INFO 02-03 01:41:32 [parallel_state.py:1423] rank 0 in world size 1 is assigned as DP rank 0, PP rank 0, PCP rank 0, TP rank 0, EP rank N/A
|
| 17 |
+
[0;36m(EngineCore_DP0 pid=3315388)[0;0m INFO 02-03 01:41:33 [gpu_model_runner.py:4021] Starting to load model Elfsong/VLM_stage_2_iter_0007000...
|
| 18 |
+
[0;36m(EngineCore_DP0 pid=3315388)[0;0m INFO 02-03 01:41:34 [cuda.py:364] Using FLASH_ATTN attention backend out of potential backends: ('FLASH_ATTN', 'FLASHINFER', 'TRITON_ATTN', 'FLEX_ATTENTION')
|
|
|
|
|
|