repo
stringclasses
147 values
number
int64
1
172k
title
stringlengths
2
476
body
stringlengths
0
5k
url
stringlengths
39
70
state
stringclasses
2 values
labels
listlengths
0
9
created_at
timestamp[ns, tz=UTC]date
2017-01-18 18:50:08
2026-01-06 07:33:18
updated_at
timestamp[ns, tz=UTC]date
2017-01-18 19:20:07
2026-01-06 08:03:39
comments
int64
0
58
user
stringlengths
2
28
vllm-project/vllm
29,865
[Doc]:
### 📚 The doc issue # Installation des bibliothèques XAI !pip install shap !pip install lime !pip install alibi !pip install interpret !pip install dalex !pip install eli5 ### Suggest a potential alternative/fix # Installation des bibliothèques XAI !pip install shap !pip install lime !pip install alibi !pip install interpret !pip install dalex !pip install eli5 ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29865
closed
[ "documentation" ]
2025-12-02T10:43:01Z
2025-12-02T10:50:00Z
0
hassaballahmahamatahmat5-cpu
vllm-project/vllm
29,864
[Usage]: I am unable to run the GLM-4.5-Air-REAP-82B-A12B-nvfp4 model on an RTX 5090.
### Your current environment I am unable to run the GLM-4.5-Air-REAP-82B-A12B-nvfp4 model on an RTX 5090. ```text Collecting environment information... ============================== System Info ============================== OS : Ubuntu 22.04.5 LTS (x86_64) GCC version : (Ubuntu 12.3.0-1ubuntu1~22.04.2) 12.3.0 Clang version : Could not collect CMake version : version 4.2.0 Libc version : glibc-2.35 ============================== PyTorch Info ============================== PyTorch version : 2.10.0.dev20251124+cu128 Is debug build : False CUDA used to build PyTorch : 12.8 ROCM used to build PyTorch : N/A ============================== Python Environment ============================== Python version : 3.10.12 (main, Nov 4 2025, 08:48:33) [GCC 11.4.0] (64-bit runtime) Python platform : Linux-6.8.0-87-generic-x86_64-with-glibc2.35 ============================== CUDA / GPU Info ============================== Is CUDA available : True CUDA runtime version : 12.8.93 CUDA_MODULE_LOADING set to : GPU models and configuration : GPU 0: NVIDIA GeForce RTX 5090 GPU 1: NVIDIA GeForce RTX 5090 GPU 2: NVIDIA GeForce RTX 5090 GPU 3: NVIDIA GeForce RTX 5090 Nvidia driver version : 570.172.08 cuDNN version : Could not collect HIP runtime version : N/A MIOpen runtime version : N/A Is XNNPACK available : True ============================== CPU Info ============================== Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 46 bits physical, 57 bits virtual Byte Order: Little Endian CPU(s): 112 On-line CPU(s) list: 0-111 Vendor ID: GenuineIntel Model name: Intel(R) Xeon(R) Gold 6330 CPU @ 2.00GHz CPU family: 6 Model: 106 Thread(s) per core: 2 Core(s) per socket: 28 Socket(s): 2 Stepping: 6 CPU max MHz: 3100.0000 CPU min MHz: 800.0000 BogoMIPS: 4000.00 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb cat_l3 intel_ppin ssbd mba ibrs ibpb stibp ibrs_enhanced tpr_shadow flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid cqm rdt_a avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb intel_pt avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local split_lock_detect wbnoinvd dtherm ida arat pln pts hwp hwp_act_window hwp_epp hwp_pkg_req vnmi avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg tme avx512_vpopcntdq la57 rdpid fsrm md_clear pconfig flush_l1d arch_capabilities Virtualization: VT-x L1d cache: 2.6 MiB (56 instances) L1i cache: 1.8 MiB (56 instances) L2 cache: 70 MiB (56 instances) L3 cache: 84 MiB (2 instances) NUMA node(s): 2 NUMA node0 CPU(s): 0-27,56-83 NUMA node1 CPU(s): 28-55,84-111 Vulnerability Gather data sampling: Mitigation; Microcode Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Mitigation; Clear CPU buffers; SMT vulnerable Vulnerability Reg file data sampling: Not affected Vulnerability Retbleed: Not affected Vulnerability Spec rstack overflow: Not affected Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; Enhanced / Automatic IBRS; IBPB conditional; RSB filling; PBRSB-eIBRS SW sequence; BHI SW loop, KVM SW loop Vulnerability Srbds: Not affected Vulnerability Tsx async abort: Not affected Vulnerability Vmscape: Not affected
https://github.com/vllm-project/vllm/issues/29864
open
[ "usage" ]
2025-12-02T10:13:31Z
2025-12-05T17:06:30Z
2
east612-ai
huggingface/diffusers
12,772
How to convert diffusers model to wan2.2 format
I see convert_wan_to_diffusers.py in diffusers repo, but no convert_diffusers_to_wan.py. Do you have plan to upload a convert scripts?
https://github.com/huggingface/diffusers/issues/12772
open
[]
2025-12-02T09:19:29Z
2025-12-02T09:19:29Z
null
wikiwen
huggingface/diffusers
12,764
When will the img2img pipeline of FLUX.2-dev be released?
I see that the current version(0.36.0-dev) only updated the text-to-image pipeline for Flux2. We are looking forward to the update of the image-to-image pipeline!
https://github.com/huggingface/diffusers/issues/12764
open
[]
2025-12-01T11:25:35Z
2025-12-01T11:41:56Z
1
guanxyu
huggingface/smolagents
1,890
Question: how to use sever-side tools provided by Google Gemini or OpenAI GPT?
Gemini has some server-side tools like google_search (https://ai.google.dev/gemini-api/docs/google-search) or google_map. OpenAI also has server-side tools like web_search. Does Smolagents support using such server-side tools from agents? If so, how?
https://github.com/huggingface/smolagents/issues/1890
open
[]
2025-12-01T05:16:01Z
2025-12-23T10:49:45Z
null
victorx-deckard
huggingface/agents-course
623
Message: Submission received, but no valid/matching task IDs were found in the 1 answers provided. Score did not improve previous record, leaderboard not updated.
I am correctly downloading the GAIA 2023 Level 1 validation dataset using snapshot_download and load_dataset. This submission is for Unit 4 Agent Course. data_dir = snapshot_download( repo_id="gaia-benchmark/GAIA", repo_type="dataset" ) dataset = load_dataset(data_dir, "2023_level1", split="validation") subset = dataset.select(range(20)) for item in subset: task_id = item.get("task_id") question_text = item.get("Question") file_name = item.get("file_name") I experience failures when trying to run the first 20 questions i received only 5 task ids are valid.. When I specifically tried to isolate and run the task ID '935e2cff-ae78-4218-b3f5-115589b19dae' using the filtering method, the evaluation system reported. <img width="1388" height="668" alt="Image" src="https://github.com/user-attachments/assets/f4e9fe1b-8608-4ab4-84cc-1b196a601694" /> 'Submission received, but no valid/matching task IDs were found in the 1 answers provided.' This occurred even though I was confident the answer was correct
https://github.com/huggingface/agents-course/issues/623
open
[ "question" ]
2025-12-01T02:09:21Z
2025-12-01T02:09:21Z
null
ShwetaBorole
huggingface/tokenizers
1,902
Guide: Compiling `tokenizers` on Android/Termux
Hello Hugging Face team and fellow developers, This is a guide for anyone trying to install `tokenizers` (or packages that depend on it, like `transformers` or `docling`) on an Android device using [Termux](https://termux.dev/). Currently, there are no other issues mentioning Termux, so hopefully, this guide can help others. ### The Problem When running `pip install tokenizers` in a standard Termux environment, the installation fails during the compilation of a C++ dependency with an error similar to this: ``` error: use of undeclared identifier 'pthread_cond_clockwait' ``` This happens because the build system is targeting an Android API level where this function is not available in the C library headers. ### The Solution The solution is to force the compilation from source and pass specific flags to the C++ compiler to set the correct Android API level and link the required libraries. Here is a step-by-step guide: #### Step 1: Install Build Dependencies You will need the Rust toolchain and other build essentials. You can install them in Termux using `pkg`: ```bash pkg update && pkg install rust clang make maturin ``` #### Step 2: Find Your Android API Level The fix requires telling the compiler which Android API level you are using. You can get this number by running the following command in your Termux shell: ```bash getprop ro.build.version.sdk ``` This will return a number, for example `29`, `30`, `33`, etc. This function (`pthread_cond_clockwait`) was introduced in API level 21, so your device's level should be higher than that. #### Step 3: Compile and Install `tokenizers` Now, you can install the package using `pip`. The command below will automatically use the API level from the previous step. ```bash # This command automatically gets your API level and uses it to compile tokenizers ANDROID_API_LEVEL=$(getprop ro.build.version.sdk) CXXFLAGS="-lpthread -D__ANDROID_API__=${ANDROID_API_LEVEL}" pip install tokenizers --no-binary :all: ``` After this, `pip install tokenizers` (and packages that depend on it) should succeed. #### Explanation of the Flags: * `CXXFLAGS="..."`: This sets environment variables to pass flags to the C++ compiler. * `-lpthread`: This flag explicitly tells the linker to link against the POSIX threads library. * `-D__ANDROID_API__=${ANDROID_API_LEVEL}`: This is the critical part. It defines a macro that tells the C++ headers to expose functions available for your specific Android version, making `pthread_cond_clockwait` visible to the compiler. * `--no-binary :all:`: This forces `pip` to ignore pre-compiled wheels and build the package from the source code, which is necessary for the flags to be applied. Hope this helps other developers working in the Termux environment!
https://github.com/huggingface/tokenizers/issues/1902
open
[]
2025-12-01T00:46:42Z
2025-12-01T00:46:42Z
0
Manamama-Gemini-Cloud-AI-01
vllm-project/vllm
29,747
[Bug]: --scheduling-policy=priority & n>1 crashes engine
### Your current environment <details> <summary>The output of <code>python collect_env.py</code></summary> ```text Your output of `python collect_env.py` here ``` </details> ### 🐛 Describe the bug When running with priority scheduling, e.g.: ```bash vllm serve Qwen/Qwen3-0.6B --scheduling-policy=priority ``` and using `n` > 1 in the request, like: ```python from openai import OpenAI client = OpenAI(base_url="http://localhost:8000/v1", api_key="dummy") res = client.chat.completions.create( model=client.models.list().data[0].id, messages=[{"role": "user", "content": "What is the meaning of life?"}], n=2 ) print(res) ``` vllm crashes with: ```python (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] EngineCore encountered a fatal error. (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] Traceback (most recent call last): (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] File "/home/user/code/debug/.venv/lib/python3.10/site-packages/vllm/v1/engine/core.py", line 835, in run_engine_core (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] engine_core.run_busy_loop() (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] File "/home/user/code/debug/.venv/lib/python3.10/site-packages/vllm/v1/engine/core.py", line 860, in run_busy_loop (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] self._process_input_queue() (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] File "/home/user/code/debug/.venv/lib/python3.10/site-packages/vllm/v1/engine/core.py", line 885, in _process_input_queue (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] self._handle_client_request(*req) (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] File "/home/user/code/debug/.venv/lib/python3.10/site-packages/vllm/v1/engine/core.py", line 907, in _handle_client_request (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] self.add_request(req, request_wave) (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] File "/home/user/code/debug/.venv/lib/python3.10/site-packages/vllm/v1/engine/core.py", line 291, in add_request (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] self.scheduler.add_request(request) (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] File "/home/user/code/debug/.venv/lib/python3.10/site-packages/vllm/v1/core/sched/scheduler.py", line 1242, in add_request (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] self.waiting.add_request(request) (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] File "/home/user/code/debug/.venv/lib/python3.10/site-packages/vllm/v1/core/sched/request_queue.py", line 150, in add_request (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] heapq.heappush(self._heap, (request.priority, request.arrival_time, request)) (EngineCore_DP0 pid=207394) ERROR 11-30 15:14:29 [core.py:844] TypeError: '<' not supported between instances of 'Request' and 'Request' (EngineCore_DP0 pid=207394) Process EngineCore_DP0: (APIServer pid=207278) ERROR 11-30 15:14:29 [async_llm.py:525] AsyncLLM output_handler failed. (APIServer pid=207278) ERROR 11-30 15:14:29 [async_llm.py:525] Traceback (most recent call last): (APIServer pid=207278) ERROR 11-30 15:14:29 [async_llm.py:525] File "/home/user/code/debug/.venv/lib/python3.10/site-packages/vllm/v1/engine/async_llm.py", line 477, in output_handler (APIServer pid=207278) ERROR 11-30 15:14:29 [async_llm.py:525] outputs = await engine_core.get_output_async() (APIServer pid=207278) ERROR 11-30 15:14:29 [async_llm.py:525] File "/home/user/code/debug/.venv/lib/python3.10/site-packages/vllm/v1/engine/core_client.py", line 883, in get_output_async (APIServer pid=207278) ERROR 11-30 15:14:29 [async_llm.py:525] raise self._format_exception(outputs) from None (APIServer pid=207278) ERROR 11-30 15:14:29 [async_llm.py:525] vllm.v1.engine.exceptions.EngineDeadError: EngineCore encountered an issue. See stack trace (above) for the root cause. (EngineCore_DP0 pid=207394) Traceback (most recent call last): (EngineCore_DP0 pid=207394) File "/usr/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap (EngineCore_DP0 pid=207394) self.run() (EngineCore_DP0 pid=207394) File "/usr/lib/python3.10/multiprocessing/process.py", line 108, in run (EngineCore_DP0 pid=207394) self._target(*self._args, **self._kwargs) (EngineCore_DP0 pid=207394) File "/home/user/code/debug/.venv/lib/python3.10/site-packages/vllm/v1/engine/core.py", line 846, in run_engine_core (EngineCore_DP0 pid=207394) raise e (EngineCore_DP0 pid=207394) File "/home/user/code/debug/.venv/lib/python3.10/site-packages/vllm/v1/engine/core.py", line 835, in run_engine_core (EngineCore_DP0 pid=207394) engine_core.run_busy_loop() (EngineCore_DP0 pid=207394) File "/home/user/code/debug/.venv/lib/python3.10/site-packages/vllm/v1/engine/core.py", line 860, in
https://github.com/vllm-project/vllm/issues/29747
closed
[ "bug" ]
2025-11-30T13:20:23Z
2025-12-02T22:42:30Z
3
hibukipanim
vllm-project/vllm
29,735
[Usage]:Accessing free_blocks count from LLMEngine or LLM ?
### Your current environment ```text None ``` ### How would you like to use vllm I'm doing research on key-value caching optimization. I want to know how to determine the number of free blocks during runtime. I tried manually creating the engine, but I couldn't find the method after searching through the code. AI keeps providing methods that have already been abandoned. I would be very grateful for any help, as this has been puzzling me for hours. ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29735
closed
[ "usage" ]
2025-11-29T19:21:50Z
2025-12-05T14:01:42Z
4
H-T-H
vllm-project/vllm
29,722
[RFC]: Add Balance Scheduling
### Motivation. **Limitations of the current vLLM v1 scheduling strategy** vLLM v1 scheduling currently enables chunkedprefill by default, which processes prefill and decode requests simultaneously in a single scheduling session. This can impact the overall system throughput and performance in some scenarios. Balance scheduling addresses this issue by synchronizing the number of running queues across all schedulers to delay the scheduling of new requests, thereby improving the overall system's steady-state decoding time. This achieves: ✅Adding `balance_gather` to the scheduler synchronizes the number of requests in the running queues between DPs. ✅Balance scheduling improves the decode steady-state time, thereby increasing the overall output throughput of the inference system. ### Proposed Change. **1.Feature Overview** In the vLLM scheduler, running requests (i.e., requests that are already undergoing pre-filled computation) have the highest priority, followed by waiting requests (i.e., requests that have not yet been computed). As shown in the diagram above, when the entire inference system exits from a steady state, the scheduler will schedule a batch of new requests for prefill operations and then synchronize them among the dynamic programming (DP) models. This can cause some DP models that are entirely decoded to synchronize with the number of prefilled tokens. Frequent prefill scheduling by certain DP models can lead to a deterioration in the overall system output throughput. Balance scheduling synchronizes the number of running queue requests across different DPs, and only schedules new requests for prefilling when at least every scheduler has fewer than max_nun_requst. **2.Implementation Design** **3.Experiment Results** - Fixed-length input scenario: In the performance test scenario with 3.5K fixed-length input and 1.5K fixed-length output, the throughput performance was improved by approximately **18%** after adding balance scheduling. | Method | Model | Input Len | Request Count | Output Len | BatchSize | Average TTFT | Average TPOT | e2e duration | Input Token Throughput | Output Token Throughput | Request Throughput | ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- | | Baseline | DeepSeekV3.1 | 3500 | 512 | 1500 | 128 | 6600 | 86.85 | 591.9s | 3030.5 | 1297.3 | 0.86 | | Balance scheduling | DeepSeekV3.1 | 3500 | 512 | 1500 | 128 | 7012 | 70.63 | 501.7s | 3575.7 | 1530.7 | 1.02 | **4.Demo PR** [#29721 ](https://github.com/vllm-project/vllm/pull/29721) ### Feedback Period. No response ### CC List. No response ### Any Other Things. No response ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29722
open
[ "RFC" ]
2025-11-29T09:28:43Z
2025-12-02T08:23:33Z
0
GDzhu01
vllm-project/vllm
29,707
[Usage]: Workaround to run model on GPUs with Compute Capability < 8.0?
### Your current environment Problem: I am unable to run the Qwen3-VL-32B-Instruct-AWQ-4bit model due to a CUDA compute capability requirement. My hardware consists of two NVIDIA QUADRO RTX 5000 cards (16GB each, 32GB total) with a compute capability of 7.5. The software framework (likely a recent version of PyTorch or a specific library) raises an error: "GPUs with compute capability < 8.0 are not supported." Question: Are there any workarounds to run this model on my older QUADRO RTX 5000 GPUs? Thanks in advance. ``` vllm collect-env INFO 11-29 20:49:15 [__init__.py:216] Automatically detected platform cuda. Collecting environment information... ============================== System Info ============================== OS : Ubuntu 24.04.3 LTS (x86_64) GCC version : (Ubuntu 13.3.0-6ubuntu2~24.04) 13.3.0 Clang version : Could not collect CMake version : version 3.30.3 Libc version : glibc-2.39 ============================== PyTorch Info ============================== PyTorch version : 2.8.0+cu128 Is debug build : False CUDA used to build PyTorch : 12.8 ROCM used to build PyTorch : N/A ============================== Python Environment ============================== Python version : 3.12.11 | packaged by Anaconda, Inc. | (main, Jun 5 2025, 13:09:17) [GCC 11.2.0] (64-bit runtime) Python platform : Linux-6.14.0-27-generic-x86_64-with-glibc2.39 ============================== CUDA / GPU Info ============================== Is CUDA available : True CUDA runtime version : 12.0.140 CUDA_MODULE_LOADING set to : LAZY GPU models and configuration : GPU 0: Quadro RTX 5000 GPU 1: Quadro RTX 5000 Nvidia driver version : 580.65.06 cuDNN version : Could not collect HIP runtime version : N/A MIOpen runtime version : N/A Is XNNPACK available : True ============================== CPU Info ============================== Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 46 bits physical, 48 bits virtual Byte Order: Little Endian CPU(s): 20 On-line CPU(s) list: 0-19 Vendor ID: GenuineIntel Model name: Intel(R) Core(TM) i9-10900X CPU @ 3.70GHz CPU family: 6 Model: 85 Thread(s) per core: 2 Core(s) per socket: 10 Socket(s): 1 Stepping: 7 CPU(s) scaling MHz: 28% CPU max MHz: 4700.0000 CPU min MHz: 1200.0000 BogoMIPS: 7399.70 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf pni pclmulqdq dtes64 monitor ds_cpl est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb cat_l3 cdp_l3 ssbd mba ibrs ibpb stibp ibrs_enhanced fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid cqm mpx rdt_a avx512f avx512dq rdseed adx smap clflushopt clwb intel_pt avx512cd avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local dtherm ida arat pln pts hwp hwp_act_window hwp_epp hwp_pkg_req avx512_vnni md_clear flush_l1d arch_capabilities L1d cache: 320 KiB (10 instances) L1i cache: 320 KiB (10 instances) L2 cache: 10 MiB (10 instances) L3 cache: 19.3 MiB (1 instance) NUMA node(s): 1 NUMA node0 CPU(s): 0-19 Vulnerability Gather data sampling: Vulnerable Vulnerability Ghostwrite: Not affected Vulnerability Itlb multihit: KVM: Mitigation: VMX unsupported Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Mitigation; Clear CPU buffers; SMT vulnerable Vulnerability Reg file data sampling: Not affected Vulnerability Retbleed: Mitigation; Enhanced IBRS Vulnerability Spec rstack overflow: Not affected Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; Enhanced / Aut
https://github.com/vllm-project/vllm/issues/29707
closed
[ "usage" ]
2025-11-29T00:47:39Z
2025-11-30T06:04:29Z
5
seasoncool
vllm-project/vllm
29,679
[Usage]: Get request total time
### Your current environment ```text ============================== System Info ============================== OS : Ubuntu 22.04.5 LTS (x86_64) GCC version : (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 Clang version : Could not collect CMake version : version 3.28.0 Libc version : glibc-2.35 ============================== PyTorch Info ============================== PyTorch version : 2.9.0+cu128 Is debug build : False CUDA used to build PyTorch : 12.8 ROCM used to build PyTorch : N/A ============================== Python Environment ============================== Python version : 3.10.19 | packaged by conda-forge | (main, Oct 22 2025, 22:29:10) [GCC 14.3.0] (64-bit runtime) Python platform : Linux-6.8.0-1030-azure-x86_64-with-glibc2.35 ============================== CUDA / GPU Info ============================== Is CUDA available : True CUDA runtime version : Could not collect CUDA_MODULE_LOADING set to : GPU models and configuration : GPU 0: NVIDIA H100 NVL Nvidia driver version : 535.247.01 cuDNN version : Probably one of the following: /usr/lib/x86_64-linux-gnu/libcudnn.so.9.10.2 /usr/lib/x86_64-linux-gnu/libcudnn_adv.so.9.10.2 /usr/lib/x86_64-linux-gnu/libcudnn_cnn.so.9.10.2 /usr/lib/x86_64-linux-gnu/libcudnn_engines_precompiled.so.9.10.2 /usr/lib/x86_64-linux-gnu/libcudnn_engines_runtime_compiled.so.9.10.2 /usr/lib/x86_64-linux-gnu/libcudnn_graph.so.9.10.2 /usr/lib/x86_64-linux-gnu/libcudnn_heuristic.so.9.10.2 /usr/lib/x86_64-linux-gnu/libcudnn_ops.so.9.10.2 HIP runtime version : N/A MIOpen runtime version : N/A Is XNNPACK available : True ============================== CPU Info ============================== Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 48 bits physical, 48 bits virtual Byte Order: Little Endian CPU(s): 40 On-line CPU(s) list: 0-39 Vendor ID: AuthenticAMD Model name: AMD EPYC 9V84 96-Core Processor CPU family: 25 Model: 17 Thread(s) per core: 1 Core(s) per socket: 40 Socket(s): 1 Stepping: 1 BogoMIPS: 4800.05 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl tsc_reliable nonstop_tsc cpuid extd_apicid aperfmperf tsc_known_freq pni pclmulqdq ssse3 fma cx16 pcid sse4_1 sse4_2 movbe popcnt aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw topoext perfctr_core vmmcall fsgsbase bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves user_shstk avx512_bf16 clzero xsaveerptr rdpru arat avx512vbmi umip avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq rdpid fsrm Hypervisor vendor: Microsoft Virtualization type: full L1d cache: 1.3 MiB (40 instances) L1i cache: 1.3 MiB (40 instances) L2 cache: 40 MiB (40 instances) L3 cache: 160 MiB (5 instances) NUMA node(s): 1 NUMA node0 CPU(s): 0-39 Vulnerability Gather data sampling: Not affected Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Not affected Vulnerability Reg file data sampling: Not affected Vulnerability Retbleed: Not affected Vulnerability Spec rstack overflow: Vulnerable: Safe RET, no microcode Vulnerability Spec store bypass: Vulnerable Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; Retpolines; STIBP disabled; RSB filling; PBRSB-eIBRS Not affected; BHI Not affected Vulnerability Srbds: Not affected Vulnerability Tsx async abort: Not affected ============================== Versions of relevant libraries ============================== [pip3] flashinfer-python==0.5.2 [pip3] numpy==1.26.4 [pip3] nvidia-cublas-cu12==12.8.4.1 [pip3] nvidia-cuda-cupti-cu12==12.8.90 [pip3] nvidia-cuda-nvrtc-cu12==12.8.93 [pip3] nvidia-cuda-runtime-cu12==12.8.90
https://github.com/vllm-project/vllm/issues/29679
closed
[ "usage" ]
2025-11-28T14:03:16Z
2025-12-01T09:34:12Z
5
chwundermsft
huggingface/lerobot
2,543
Different finetune loss given policy.type=pi0 / policy.path=lerobot/pi0_base. What is the difference?
Hi, I have two different configurations: 1. ` --dataset.repo_id=BBBBBBob/libero_goal_lerobot \ --dataset.root=/home/j84403411/data/libero/libero_goal_lerobot \ --policy.path=lerobot/pi0_base \ --policy.push_to_hub=false \ --policy.use_proprio=true \ --output_dir=/home/j84403411/checkpoint/libero/pi0/libero_goal_proprio \ --policy.dtype=bfloat16 \ --steps=40_000 \ --batch_size=16 \ --rename_map='{"observation.images.image":"observation.images.base_0_rgb", "observation.images.wrist_image":"observation.images.left_wrist_0_rgb"}' \ ` and 2. ` --dataset.repo_id=BBBBBBob/libero_goal_lerobot \ --dataset.root=/home/j84403411/data/libero/libero_goal_lerobot \ --policy.type=pi0 \ --policy.pretrained_path=lerobot/pi0_base \ --policy.push_to_hub=false \ --policy.use_proprio=true \ --output_dir=/home/j84403411/checkpoint/libero/pi0/libero_goal_proprio \ --policy.dtype=bfloat16 \ --steps=40_000 \ --batch_size=16 \ --policy.input_features='{"observation.state": {"type": "STATE", "shape": [8]}, "observation.images.wrist_image": {"type": "VISUAL", "shape": [3, 256, 256]}, "observation.images.image": {"type": "VISUAL", "shape": [3, 256, 256]}, }' \ --policy.output_features='{"action": {"type": "ACTION", "shape": [7]}}' \ ` The loss trained from the second configuration is 10 times higher than the first one. What caused the difference? Do you know if different checkpoints are loaded in this case? I appreciate your help!
https://github.com/huggingface/lerobot/issues/2543
closed
[]
2025-11-28T12:34:38Z
2025-12-01T11:25:17Z
null
BBBBBBob
huggingface/transformers.js
1,467
Missing the following inputs: input_points, input_labels (or input_boxes)
### Question thanks for your excellent works! I just write test code for SlimSAM model powered by transformers.js referring to this example(with some improvements): https://github.com/huggingface/transformers.js-examples/blob/main/segment-anything-webgpu/index.js my code for `decode` method: ```js // Decode segmentation async function decode() { if (!imageEmbeddings || isDecoding || isEncoding) return; if (isDecoding) { decodePending = true; return; } isDecoding = true; try { let input_points = null; let input_labels = null; let input_boxes = null; let outputs = null; if (promptMode == "point" && points.length > 0) { const reshaped = imageprocessed.reshaped_input_sizes[0]; // [H, W] const scaledPoints = points.map(p => [ p.x * reshaped[1], p.y * reshaped[0] ]); const labels = points.map(p => BigInt(p.label)); input_points = new Tensor("float32", scaledPoints.flat(), [1, 1, points.length, 2]); input_labels = new Tensor("int64", labels, [1, 1, points.length]); // Fallback: if no prompts, skip if (!input_points) return; // Run model with point mode outputs = await model({ ...imageEmbeddings, input_points: input_points, input_labels: input_labels, input_boxes: null }); } if (promptMode == "box" && box) { const reshaped = imageprocessed.reshaped_input_sizes[0]; const [x1, y1, x2, y2] = [ box.x1 * reshaped[1], box.y1 * reshaped[0], box.x2 * reshaped[1], box.y2 * reshaped[0] ]; input_boxes = new Tensor("float32", [x1, y1, x2, y2], [1, 1, 4]); // Fallback: if no prompts, skip if (!input_boxes) return; // Run model with box mode outputs = await model({ ...imageEmbeddings, input_points: null, input_labels: null, input_boxes: input_boxes }); } // Post-process const masks = await processor.post_process_masks( outputs.pred_masks, imageprocessed.original_sizes, imageprocessed.reshaped_input_sizes ); const scores = outputs.iou_scores.data; updateMask(masks[0], scores); // masks[0] is [3, H, W] } catch (e) { console.error("Decode error:", e); statusEl.textContent = "❌ Segmentation failed."; } finally { isDecoding = false; if (decodePending) { decodePending = false; decode(); } } } ``` it supports 2 prompt modes: `point` &` box` which selected by users on UI elements (html not provided). but error printed every time when running `decode` method (at the line of calling `outputs = await model(...)`), the error message is: with box prompt mode: `Error: An error occurred during model execution: "Missing the following inputs: input_points, input_labels.` with point prompt mode: `Error: An error occurred during model execution: "Missing the following inputs: input_boxes.` Should I pass all three parameters(input_points/input_labels/input_boxes) simultaneously, regardless of which prompt mode I’m using? How could I support point & box at the same time, since no demo codes found on internet. thanks! ``` version: transformers.js 3.5.0 from https://cdn.jsdelivr.net/npm/@huggingface/transformers@3.5.0 os: Windows 10 chorme: 142 model: Xenova/slimsam-77-uniform ```
https://github.com/huggingface/transformers.js/issues/1467
closed
[ "question" ]
2025-11-28T10:01:04Z
2025-12-01T04:04:59Z
null
sherlockchou86
vllm-project/vllm
29,643
[Usage]: Enabling Tool call in the Python SDK
### Your current environment Hi Team, I am currently exploring VLLM to enable tool calling, and I need some support with this. It would be very helpful if you could provide the corresponding Python code. What I’m trying to achieve is to configure the Python package with the same settings that I use when starting the VLLM server. The configuration I’m using is: vllm serve DeepSeek-R1-0528-Qwen3-8B \ --served-model-name deepseek \ --gpu_memory_utilization 0.5 \ --max_num_seqs 20 \ --max_model_len 10000 \ --enable-auto-tool-choice \ --tool-call-parser deepseek_v3 \ --chat-template tool_chat_template_deepseekr1.jinja \ --port 5050 \ --max_num_batched_tokens 5000 I need to replicate this exact configuration in Python. Your support would be greatly appreciated. Please respond at your earliest convenience. If you want, I can also write the **Python code equivalent** for these VLLM configurations. Best Regards Madan ### How would you like to use vllm I want to use vLLM to serve a model with tool-calling support enabled. Specifically, I need to run the model with the same configuration parameters that I currently use when launching the vLLM server from the command line. These settings include GPU memory utilization, maximum sequence limits, tool-calling options, a custom tool-call parser, and a custom chat template. My goal is to reproduce the following server configuration within a Python environment using the vLLM Python API: vllm serve DeepSeek-R1-0528-Qwen3-8B \ --served-model-name deepseek \ --gpu_memory_utilization 0.5 \ --max_num_seqs 20 \ --max_model_len 10000 \ --enable-auto-tool-choice \ --tool-call-parser deepseek_v3 \ --chat-template tool_chat_template_deepseekr1.jinja \ --port 5050 \ --max_num_batched_tokens 5000 ` In short, I need Python code that sets these exact configurations so I can run vLLM programmatically with tool calling enabled. If you want, I can also provide the **Python code equivalent** for this configuration. ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29643
open
[ "usage" ]
2025-11-28T04:39:47Z
2025-12-01T14:54:47Z
2
Madan1215
vllm-project/vllm
29,641
[Bug]: Max Tokens not being honoured in Chat Completions for GPTOSS model
### Your current environment It seems that in the latest version of vllm 0.11+ Chat Completions has stopped honouring `max_tokens` with GPTOSS 120B model, the below request payload has stopped working with `max_tokens` earlier the same payload would provide an output to the limit of the `max_tokens` provided.. Interestingly if you look at the `usage` tokens, it's showing `completion_tokens` as 500 but the output is BLANK. ```json { "messages": [ { "role": "user", "content": "What is the role of AI in medicine?" } ], "model": "openai/gpt-oss-120b", "max_tokens": 500, "reasoning": {"effort": "low"}, "stream": false } ``` getting BLANK output, even though the `usage` is showing token counts created is matching max_tokens ```json { "id": "chatcmpl-c71e934ac0b74bd4b8f99fe9b5516ea3", "object": "chat.completion", "created": 1764300020, "model": "openai/gpt-oss-120b", "choices": [ { "index": 0, "message": { "role": "assistant", "content": null, "refusal": null, "annotations": null, "audio": null, "function_call": null, "tool_calls": [], "reasoning": "Need to answer.", "reasoning_content": "Need to answer." }, "logprobs": null, "finish_reason": "length", "stop_reason": null, "token_ids": null } ], "service_tier": null, "system_fingerprint": null, "usage": { "prompt_tokens": 78, "total_tokens": 578, "completion_tokens": 500, "prompt_tokens_details": null }, "prompt_logprobs": null, "prompt_token_ids": null, "kv_transfer_params": null } ``` When you remove the `max_tokens`, we get the output which shows `usage_token` to have `completion_tokens` to be around 1600 tokens.. It seems that starting from vllm 0.11+ version, the auto-truncation using the `max_tokens` has stopped working ```json { "id": "chatcmpl-61b60144d43147e2b007158712ad4920", "object": "chat.completion", "created": 1764300423, "model": "openai/gpt-oss-120b", "choices": [ { "index": 0, "message": { "role": "assistant", "content": "**The role of AI in medicine is expanding rapidly and touches virtually every aspect of healthcare—from the way doctors diagnose patients to how hospitals run their operations.** Below is a structured overview that covers the major domains, concrete examples, benefits, challenges, and future directions.\n\n---\n\n## 1. Clinical Care\n\n| Sub‑area | What AI Does | Real‑World Examples | Benefits |\n|----------|--------------|---------------------|----------|\n| **Diagnostics** | Image analysis, pattern recognition, risk stratification | • Radiology: Google DeepMind’s AI detects lung cancer on CT scans with >95% accuracy.<br>• Dermatology: FDA‑cleared apps (e.g., SkinVision) classify skin lesions from photos.<br>• Pathology: Paige.ai assists in detecting prostate cancer in biopsy slides. | Faster, more consistent readings; can catch subtle findings that human eyes miss. |\n| **Predictive Analytics** | Forecast disease onset, complications, readmission risk | • Sepsis prediction models (e.g., Epic Sepsis Model) trigger alerts hours before clinical signs.<br>• Cardiovascular risk calculators incorporating genomics and wearables. | Enables proactive interventions, reduces morbidity and cost. |\n| **Treatment Planning** | Decision support, dose optimisation, drug selection | • IBM Watson for Oncology (clinical trial matching).<br>• Radiation oncology: AI‑driven dose‑painting to spare healthy tissue.<br>• Pharmacogenomics: AI predicts drug‑gene interactions. | Personalises therapy, improves outcomes, reduces adverse events. |\n| **Robotics & Minimally Invasive Surgery** | Real‑time image guidance, autonomous suturing, task automation | • Da Vinci Surgical System (augmented with AI for instrument tracking).<br>• VERDICT AI for autonomous suturing in animal models. | Increases precision, reduces surgeon fatigue, shortens recovery. |\n\n---\n\n## 2. Patient‑Facing Applications\n\n| Application | Description | Example |\n|-------------|-------------|---------|\n| **Virtual Assistants & Chatbots** | Symptom triage, medication reminders, mental‑health chat | • Babylon Health (AI‑driven triage).<br>• Woebot (CBT‑based mental‑health chatbot). |\n| **Telemedicine Enhancements** | Real‑time vitals extraction from video, automated note‑taking | • KardiaMobile ECG integration with AI‑based arrhythmia detection. |\n| **Wearables & Remote Monitoring** | Continuous data streams analysed for early alerts | • Apple Watch ECG + AI arrhythmia detection; Fitbit heart‑rate trend alerts. |\n\n---\n\n## 3. Operational & Administrative Efficiency\n\n| Domain | AI Functions | Example |
https://github.com/vllm-project/vllm/issues/29641
closed
[ "bug" ]
2025-11-28T03:39:34Z
2025-12-21T02:39:32Z
16
soodrohit
huggingface/transformers
42,464
Add SAM 3D Objects Encoder
### Model description ## Model Description SAM 3D Objects is Meta AI's foundation model for 3D object reconstruction from single images. I'm proposing to add the **encoder component** (DINOv2-based Vision Transformer) to Transformers. **Scope**: Encoder only, not the full 3D generation pipeline (which includes Gaussian Splatting/Mesh decoders better suited for Diffusers). ## Open source status - [x] The model implementation available - [x] The model weights are available ## Provide useful links for the implementation - **Model Card**: https://huggingface.co/facebook/sam-3d-objects - **Paper**: https://arxiv.org/abs/2511.16624 - **Original Repository**: https://github.com/facebookresearch/sam-3d-objects - **Blog Post**: https://ai.meta.com/blog/sam-3d/ ## Implementation Progress I have already implemented this model and it's ready for review: ✅ **Implementation Complete:** - `Sam3DObjectsEncoderConfig` - Configuration with DINO variant support - `Sam3DObjectsEncoder` - Main encoder model - `Sam3DObjectsEncoderForMasks` - Variant for mask encoding - `Sam3DObjectsImageProcessor` - Image preprocessing - Comprehensive test suite: **28/28 tests passing** - Full documentation **Test Results:** collected 29 items 28 passed, 1 skipped in 4.92s **Example Usage:** ```python from transformers.models.sam3d_objects import ( Sam3DObjectsEncoder, Sam3DObjectsEncoderConfig, Sam3DObjectsImageProcessor, ) config = Sam3DObjectsEncoderConfig.from_dino_config("dinov2_vitl14") model = Sam3DObjectsEncoder(config) processor = Sam3DObjectsImageProcessor() inputs = processor(images=image, return_tensors="pt") outputs = model(**inputs) embeddings = outputs.last_hidden_state ``` ## Questions 1. Is there interest in adding the SAM 3D Objects Encoder to Transformers? 2. Should this be limited to the encoder component (my recommendation)? 3. Should I submit a PR, or are there any requirements I should address first? ## Additional Context - The encoder is based on DINOv2 and fits naturally in Transformers - Full 3D generation pipeline would be better suited for Diffusers - Model is gated on Hub (requires license acceptance) - Implementation follows Transformers patterns and guidelines I'm ready to submit a PR and address any feedback. ### Open source status - [x] The model implementation is available - [x] The model weights are available ### Provide useful links for the implementation ## Links - **Model Card**: https://huggingface.co/facebook/sam-3d-objects - **Paper**: https://arxiv.org/abs/2511.16624 (SAM 3D: 3Dfy Anything in Images) - **Original Repository**: https://github.com/facebookresearch/sam-3d-objects - **Blog Post**: https://ai.meta.com/blog/sam-3d/ - **Project Page**: https://ai.meta.com/sam3d/ ## Authors **SAM 3D Team** from Meta AI For the complete author list and contributions, see: - [ArXiv Paper](https://arxiv.org/abs/2511.16624) - [Original Repository](https://github.com/facebookresearch/sam-3d-objects) *Note: This is a large collaborative project with many contributors from Meta Superintelligence Labs.* ## Implementation Details **Model Type**: Vision Encoder (DINOv2-based) **Architecture**: Vision Transformer (ViT) **Variants Supported**: - ViT-S/14 (384 dim) - ViT-B/14 (768 dim) - ViT-L/14 (1024 dim) - ViT-G/14 (1536 dim) **Input**: RGB images (224x224 or 518x518) **Output**: Visual embeddings for 3D generation tasks **License**: SAM License (gated model on HuggingFace Hub)
https://github.com/huggingface/transformers/issues/42464
open
[ "New model" ]
2025-11-27T19:48:28Z
2025-12-05T10:32:33Z
1
Aznix07
vllm-project/vllm
29,584
[Usage]: Can KV Cache be disabled in non-autoregressive generation tasks?
### Your current environment ```text ============================== System Info ============================== OS : Ubuntu 24.04.3 LTS (x86_64) GCC version : (Ubuntu 13.3.0-6ubuntu2~24.04) 13.3.0 Clang version : Could not collect CMake version : version 3.28.3 Libc version : glibc-2.39 ============================== PyTorch Info ============================== PyTorch version : 2.9.0+cu128 Is debug build : False CUDA used to build PyTorch : 12.8 ROCM used to build PyTorch : N/A ============================== Python Environment ============================== Python version : 3.10.15 (main, Oct 3 2024, 07:27:34) [GCC 11.2.0] (64-bit runtime) Python platform : Linux-6.8.0-87-generic-x86_64-with-glibc2.39 ============================== CUDA / GPU Info ============================== Is CUDA available : True CUDA runtime version : Could not collect CUDA_MODULE_LOADING set to : GPU models and configuration : GPU 0: NVIDIA GeForce RTX 3090 GPU 1: NVIDIA GeForce RTX 3090 Nvidia driver version : 575.57.08 cuDNN version : Could not collect HIP runtime version : N/A MIOpen runtime version : N/A Is XNNPACK available : True ============================== vLLM Info ============================== ROCM Version : Could not collect vLLM Version : 0.11.2 vLLM Build Flags: CUDA Archs: Not Set; ROCm: Disabled GPU Topology: GPU0 GPU1 CPU Affinity NUMA Affinity GPU NUMA ID GPU0 X SYS 0-23,48-71 0 N/A GPU1 SYS X 24-47,72-95 1 N/A Legend: X = Self SYS = Connection traversing PCIe as well as the SMP interconnect between NUMA nodes (e.g., QPI/UPI) NODE = Connection traversing PCIe as well as the interconnect between PCIe Host Bridges within a NUMA node PHB = Connection traversing PCIe as well as a PCIe Host Bridge (typically the CPU) PXB = Connection traversing multiple PCIe bridges (without traversing the PCIe Host Bridge) PIX = Connection traversing at most a single PCIe bridge NV# = Connection traversing a bonded set of # NVLinks ``` ### How would you like to use vllm Hello vLLM team, Currently, vLLM (v0.11.2) enables KV cache for certain LLM-based pooling and reranking models, such as the Qwen3-Embedding series, even when `--no-enable-chunked-prefill` and `--no-enable-prefix-caching` are set. This leads to unnecessary GPU memory usage. Would it be possible to disable KV cache for pooling and reranking models under these conditions? ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29584
open
[ "usage" ]
2025-11-27T05:30:08Z
2025-12-05T02:40:28Z
5
GitEventhandler
vllm-project/vllm
29,574
[Performance]: Using vLLM to accelerate VLM models, does the vision encoding part currently support parallel processing, or is it still being processed serially?
### Proposal to improve performance I found that currently, images of different sizes are processed sequentially, which significantly slows down the processing speed. How can we adapt to parallel processing? Should we resize or pad all images to the same size for batch processing, or can we run multiple encoder models in parallel? Thank you. ### Report of performance regression _No response_ ### Misc discussion on performance _No response_ ### Your current environment (if you think it is necessary) ```text The output of `python collect_env.py` ``` ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29574
open
[ "performance" ]
2025-11-27T03:51:36Z
2025-11-27T10:54:09Z
2
NewZxy
vllm-project/vllm
29,564
[Doc]: Make PyTorch profiler gzip and CUDA time dump configurable
### 📚 The doc issue We observed that enabling both use_gzip and dump_self_cuda_time_total in the vLLM torch profiler introduces significant overhead during profiling. For example, when profiling 10 randomly generated requests (1000 input tokens, 200 output tokens) on an A100 using the Qwen3-32B model, we found that gzip compression of the profiling trace and dumping the CUDA time table take ~68 seconds, dominating the overall profiling time. The main sources of overhead appear to be: 1. Gzip compression of the profiling trace file 2. Generation and dumping of the CUDA time summary table After disabling these two features, the total profiling dump time is reduced to ~18 seconds. In many profiling scenarios (e.g., quick performance checks or small-scale experiments), users may not need gzip compression or the CUDA time table. Therefore, it would be helpful to make these two behaviors individually configurable via environment variables—enabled by default for completeness, but optionally turnable off when faster profiling turnaround is preferred. Moreover, gzip compression could potentially be performed asynchronously after the trace is dumped, allowing lower-latency profiling in staging or pre-production environments. This patch proposes adding such configurability so users can selectively disable gzip compression and/or CUDA time table generation when they want a faster and lighter profiling workflow. ### Suggest a potential alternative/fix _No response_ ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29564
closed
[ "documentation" ]
2025-11-27T02:21:20Z
2025-12-01T04:30:48Z
1
zhangruoxu
vllm-project/vllm
29,562
[Bug]: "\n\n" content between reasoning and tool_call content when tool_call and stream mode
### Your current environment <details> <summary>The output of <code>python collect_env.py</code></summary> ```text Your output of `python collect_env.py` here ``` </details> ### 🐛 Describe the bug https://github.com/QwenLM/Qwen3/issues/1755 When stream mode true, the response contains content "\n\n" between reasoning and tool_call; but with stream model false, it didn't generate content "\n\n". Is there some thing different, I don't want the content "\n\n" between reasoning and tool_call. <img width="974" height="533" alt="Image" src="https://github.com/user-attachments/assets/0cc36343-3c0f-4ce1-9028-30f561a55dac" /> Here is my requests: ``` { "model": "Qwen3-235B-A22B-Thinking-2507", "tools": [ { "type": "function", "function": { "name": "search_law_articles", "parameters": { "type": "object", "properties": { "level": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "default": null, "description": "搜索条件:法规类型" }, "query": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "default": null, "description": "查询语句" }, "title": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "default": null, "description": "法律标题" }, "article": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "default": null, "description": "法律条款序号,如 第十条" }, "content": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "default": null, "description": "法律条款及内容,如 第十条 贷款人委托支付" }, "pub_department": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "default": null, "description": "发布部门" }, "pub_time_after": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "default": null, "description": "搜索条件:发布时间晚于此时间,格式如2025-06-20" }, "pub_time_before": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "default": null, "description": "搜索条件:发布时间早于此时间,格式如2025-06-20" }, "imply_time_after": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "default": null, "description": "搜索条件:实施时间晚于此时间,格式如2025-06-20" }, "imply_time_before": { "anyOf": [ { "type": "string" }, { "type": "null" } ], "default": null, "description": "搜索条件:实施时间早于此时间,格式如2025-06-20" } } }, "description": "此工具用于搜索法条内容, 库中是按照法律条目进行存储, 查询可选多个查询过滤条件" } } ], "stream": true, "messages": [ { "role": "user", "content": [ { "text": "帮我解读下网络安全法", "type": "text" } ] } ] } ``` ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29562
open
[ "bug" ]
2025-11-27T01:49:04Z
2025-11-27T01:49:04Z
0
NiuBlibing
vllm-project/vllm
29,560
[Doc]: Batch Invariance on Ampere Platforms
### 📚 The doc issue Does the batch invariance feature released in vllm 0.11.2 support the Ampere architecture? If adaptations are required, what modifications need to be made? ### Suggest a potential alternative/fix _No response_ ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29560
closed
[ "documentation" ]
2025-11-27T01:06:49Z
2025-11-27T14:21:30Z
0
luo1206
huggingface/trl
4,582
Does the GRPO Trainer support multi-image input for Qwen3-VL?
Does the GRPO Trainer support multi-image input for Qwen3-VL?
https://github.com/huggingface/trl/issues/4582
open
[ "🏋 GRPO" ]
2025-11-26T14:03:57Z
2025-11-27T08:08:25Z
1
Lestoky
huggingface/diffusers
12,722
How to run qwen-image in kaggle gpu T4 * 2 successfully?
```python3 !python3 -m pip install -U diffusers peft bitsandbytes import diffusers, torch, math qwen = diffusers.QwenImagePipeline.from_pretrained('Qwen/Qwen-Image', torch_dtype=torch.float16, low_cpu_mem_usage=True, quantization_config=diffusers.PipelineQuantizationConfig(quant_backend='bitsandbytes_4bit', quant_kwargs={'load_in_4bit':True, 'bnb_4bit_quant_type':'nf4', 'bnb_4bit_compute_dtype':torch.float16}, components_to_quantize=['transformer', 'text_encoder'])) qwen.scheduler = diffusers.FlowMatchEulerDiscreteScheduler.from_config({'base_image_seq_len':256, 'base_shift':math.log(3), 'invert_sigmas':False, 'max_image_seq_len':8192, 'max_shift':math.log(3), 'num_train_timesteps':1000, 'shift':1, 'shift_terminal':None, 'stochastic_sampling':False, 'time_shift_type':'exponential', 'use_beta_sigmas':False, 'use_dynamic_shifting':True, 'use_exponential_sigmas':False, 'use_karras_sigmas':False}) qwen.load_lora_weights('lightx2v/Qwen-Image-Lightning', weight_name='Qwen-Image-Lightning-4steps-V2.0.safetensors', adapter_name='lightning') qwen.set_adapters('lightning', adapter_weights=1) qwen.enable_sequential_cpu_offload() qwen(prompt='a beautiful girl', height=1280, width=720, num_inference_steps=4, true_cfg_scale=1).images[0].save('a.png') ``` ----> 3 qwen = diffusers.QwenImagePipeline.from_pretrained('Qwen/Qwen-Image', torch_dtype=torch.float16, low_cpu_mem_usage=True, quantization_config=diffusers.PipelineQuantizationConfig(quant_backend='bitsandbytes_4bit', quant_kwargs={'load_in_4bit':True, 'bnb_4bit_quant_type':'nf4', 'bnb_4bit_compute_dtype':torch.float16}, components_to_quantize=['transformer', 'text_encoder'])) OutOfMemoryError: CUDA out of memory. Tried to allocate 34.00 MiB. GPU 0 has a total capacity of 14.74 GiB of which 4.19 MiB is free. Process 8568 has 14.73 GiB memory in use. Of the allocated memory 14.50 GiB is allocated by PyTorch, and 129.00 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) How to get more cuda memory? @yiyixuxu @DN6
https://github.com/huggingface/diffusers/issues/12722
open
[]
2025-11-26T12:53:30Z
2025-11-28T03:54:07Z
null
chaowenguo
vllm-project/vllm
29,494
[Doc]: Documentation inconsistency: Blog mentions append_slots() but codebase uses allocate_slots()
### 📚 The doc issue The Automatic Prefix Caching blog post mentions: > "The scheduler calls kv_cache_manager.append_slots()" However, the actual codebase uses a unified `kv_cache_manager.allocate_slots()` method that handles both prefill and decode requests. **Location:** - Blog: [[link to blog post](https://docs.vllm.ai/en/v0.8.5/design/v1/prefix_caching.html#operations)] - Code: ./vllm/v1/core/kv_cache_manager.py ### Suggest a potential alternative/fix Update the blog post to reflect the actual implementation `kv_cache_manager.allocate_slots()` ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29494
closed
[ "documentation" ]
2025-11-26T11:37:40Z
2025-11-26T11:46:08Z
1
pradsgit
huggingface/transformers
42,418
Custom nn.Parameter initialization in PreTrainedModel subclasses is overwritten by post_init()/from_pretrained() causing NaNs/Zeros
### System Info - `transformers` version: 4.57.1 - Platform: Linux-4.18.0-147.mt20200626.413.el8_1.x86_64-x86_64-with-glibc2.35 - Python version: 3.10.14 - Huggingface_hub version: 0.35.3 - Safetensors version: 0.6.2 - Accelerate version: 1.11.0 - Accelerate config: not found - DeepSpeed version: 0.18.2 - PyTorch version (accelerator?): 2.7.1+cu126 (CUDA) - Tensorflow version (GPU?): not installed (NA) - Flax version (CPU?/GPU?/TPU?): not installed (NA) - Jax version: not installed - JaxLib version: not installed - Using distributed or parallel set-up in script?: No - Using GPU in script?: No - GPU type: NVIDIA A100-SXM4-80GB ### Who can help? @Cyrilvallez @zucchini-nlp ### Information - [ ] The official example scripts - [x] My own modified scripts ### Tasks - [ ] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...) - [x] My own task or dataset (give details below) ### Reproduction ```python import numpy as np import os import random import torch import torch.nn as nn from transformers import Qwen3VLForConditionalGeneration def seed_everything(TORCH_SEED): random.seed(TORCH_SEED) os.environ["PYTHONHASHSEED"] = str(TORCH_SEED) np.random.seed(TORCH_SEED) torch.manual_seed(TORCH_SEED) torch.cuda.manual_seed(TORCH_SEED) torch.cuda.manual_seed_all(TORCH_SEED) torch.backends.cudnn.deterministic = True torch.backends.cudnn.benchmark = False seed_everything(66) class TestModel1(Qwen3VLForConditionalGeneration): def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.action_head = nn.Linear(1024, 7) self.positional_embedding = nn.Parameter(torch.randn(16, 1152)) self.post_init() class TestModel2(nn.Module): def __init__(self, *args, model_path, **kwargs): super().__init__(*args, **kwargs) self.model = Qwen3VLForConditionalGeneration.from_pretrained(model_path) self.action_head = nn.Linear(1024, 7) self.positional_embedding = nn.Parameter(torch.randn(16, 1152)) test_model1 = TestModel1.from_pretrained("Qwen/Qwen3-VL-4B-Instruct") test_model2 = TestModel2(model_path="Qwen/Qwen3-VL-4B-Instruct") print(test_model1.positional_embedding) print(test_model1.positional_embedding.mean(), test_model1.positional_embedding.std()) print(test_model2.positional_embedding) print(test_model2.positional_embedding.mean(), test_model2.positional_embedding.std()) ```` ### Expected behavior When subclassing a model (inheriting from PreTrainedModel, e.g., Qwen3VLForConditionalGeneration, LlamaForCausalLM) to add custom learnable parameters, user-defined initialization in __init__ is often silently overwritten. This occurs because from_pretrained (or the end of __init__) triggers self.post_init(), which recursively calls _init_weights. This mechanism re-initializes all parameters, ignoring the explicit initialization code provided by the user in __init__. In the specific case of Qwen3-VL (and potentially others), this re-initialization results in NaNs or Zeros, rendering the model unusable without manual intervention. Steps to reproduce The following script demonstrates the issue. Note: I used torch.randn for the custom parameter initialization. While I understand that torch.randn samples from a standard normal distribution and does not guarantee an exact sample mean of 0 and std of 1, it should result in valid float values. The observed NaNs/Zeros confirm that this initialization is being discarded and replaced by a faulty internal initialization logic.
https://github.com/huggingface/transformers/issues/42418
open
[ "Usage", "Feature request", "bug" ]
2025-11-26T10:29:57Z
2025-12-01T15:10:32Z
10
Noietch
huggingface/diffusers
12,720
how to quantization wan 2.2 vace after loading lora?
```python3 diffusers.WanVACEPipeline.from_pretrained('linoyts/Wan2.2-VACE-Fun-14B-diffusers', vae=diffusers.AutoencoderKLWan.from_pretrained('linoyts/Wan2.2-VACE-Fun-14B-diffusers', subfolder='vae', torch_dtype=torch.float32), torch_dtype=torch.bfloat16, quantization_config=diffusers.PipelineQuantizationConfig(quant_backend='bitsandbytes_8bit', quant_kwargs={'load_in_8bit':True}, components_to_quantize=['transformer', 'transformer_2'])).save_pretrained('wan') ``` normally I can save the quantization model in this way But now I want to merge lora and the quantization and then save the model with lora. How? ```python3 wan = diffusers.WanVACEPipeline.from_pretrained('linoyts/Wan2.2-VACE-Fun-14B-diffusers', vae=diffusers.AutoencoderKLWan.from_pretrained('linoyts/Wan2.2-VACE-Fun-14B-diffusers', subfolder='vae', torch_dtype=torch.float32), torch_dtype=torch.bfloat16) wan.load_lora_weights('lightx2v/Wan2.2-Lightning', weight_name='Wan2.2-I2V-A14B-4steps-lora-rank64-Seko-V1/high_noise_model.safetensors', adapter_name='lightning') wan.load_lora_weights('lightx2v/Wan2.2-Lightning', weight_name='Wan2.2-I2V-A14B-4steps-lora-rank64-Seko-V1/low_noise_model.safetensors', adapter_name='lightning_2', load_into_transformer_2=True) wan.set_adapters(['lightning', 'lightning_2'], adapter_weights=[1] * 2) how to quantization and save_pretrained? ``` @yiyixuxu @DN6
https://github.com/huggingface/diffusers/issues/12720
open
[]
2025-11-26T10:11:38Z
2025-12-11T17:29:30Z
null
chaowenguo
vllm-project/vllm
29,489
[Usage]: Removing last generated token from output and kv cache
### Your current environment ```text Collecting environment information... ============================== System Info ============================== OS : Ubuntu 24.04.3 LTS (x86_64) GCC version : (Ubuntu 13.3.0-6ubuntu2~24.04) 13.3.0 Clang version : Could not collect CMake version : version 3.28.3 Libc version : glibc-2.39 ============================== PyTorch Info ============================== PyTorch version : 2.9.0+cu128 Is debug build : False CUDA used to build PyTorch : 12.8 ROCM used to build PyTorch : N/A ============================== Python Environment ============================== Python version : 3.13.5 | packaged by conda-forge | (main, Jun 16 2025, 08:27:50) [GCC 13.3.0] (64-bit runtime) Python platform : Linux-6.8.0-87-generic-x86_64-with-glibc2.39 ============================== CUDA / GPU Info ============================== Is CUDA available : True CUDA runtime version : 12.8.93 CUDA_MODULE_LOADING set to : GPU models and configuration : GPU 0: NVIDIA B200 GPU 1: NVIDIA B200 GPU 2: NVIDIA B200 GPU 3: NVIDIA B200 GPU 4: NVIDIA B200 GPU 5: NVIDIA B200 GPU 6: NVIDIA B200 GPU 7: NVIDIA B200 Nvidia driver version : 570.195.03 cuDNN version : Could not collect HIP runtime version : N/A MIOpen runtime version : N/A Is XNNPACK available : True ============================== CPU Info ============================== Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 52 bits physical, 57 bits virtual Byte Order: Little Endian CPU(s): 224 On-line CPU(s) list: 0-223 Vendor ID: GenuineIntel Model name: INTEL(R) XEON(R) PLATINUM 8570 CPU family: 6 Model: 207 Thread(s) per core: 2 Core(s) per socket: 56 Socket(s): 2 Stepping: 2 CPU(s) scaling MHz: 33% CPU max MHz: 4000.0000 CPU min MHz: 800.0000 BogoMIPS: 4200.00 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf tsc_known_freq pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb cat_l3 cat_l2 cdp_l3 intel_ppin cdp_l2 ssbd mba ibrs ibpb stibp ibrs_enhanced tpr_shadow flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid cqm rdt_a avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb intel_pt avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local split_lock_detect user_shstk avx_vnni avx512_bf16 wbnoinvd dtherm ida arat pln pts hwp hwp_act_window hwp_epp hwp_pkg_req vnmi avx512vbmi umip pku ospke waitpkg avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg tme avx512_vpopcntdq la57 rdpid bus_lock_detect cldemote movdiri movdir64b enqcmd fsrm md_clear serialize tsxldtrk pconfig arch_lbr ibt amx_bf16 avx512_fp16 amx_tile amx_int8 flush_l1d arch_capabilities ibpb_exit_to_user Virtualization: VT-x L1d cache: 5.3 MiB (112 instances) L1i cache: 3.5 MiB (112 instances) L2 cache: 224 MiB (112 instances) L3 cache: 600 MiB (2 instances) NUMA node(s): 2 NUMA node0 CPU(s): 0-55,112-167 NUMA node1 CPU(s): 56-111,168-223 Vulnerability Gather data sampling: Not affected Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Not affected Vulnerability Reg file data sampling: Not affected Vulnerability Retbleed: Not affected Vulnerability Spec rstack overflow: Not affected Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; Enhanced / Automatic IBRS; IBPB conditional; RSB filling; PBRSB-eIBRS SW sequence; BHI
https://github.com/vllm-project/vllm/issues/29489
open
[ "usage" ]
2025-11-26T09:35:37Z
2025-11-26T09:36:37Z
0
josefdra
huggingface/diffusers
12,719
how to use quantization and device_map=balance to run qwen-image on kaggle T4 * 2
```python3 !python3 -m pip install -U diffusers peft bitsandbytes protobuf import diffusers, torch, math qwen = diffusers.QwenImagePipeline.from_pretrained('Qwen/Qwen-Image', quantization_config=diffusers.PipelineQuantizationConfig(quant_backend='bitsandbytes_4bit', quant_kwargs={'load_in_4bit':True, 'bnb_4bit_quant_type':'nf4', 'bnb_4bit_compute_dtype':torch.float16}, components_to_quantize=['transformer', 'text_encoder']), torch_dtype=torch.float16, device_map='balanced') print(qwen.hf_device_map) qwen.scheduler = diffusers.FlowMatchEulerDiscreteScheduler.from_config({'base_image_seq_len':256, 'base_shift':math.log(3), 'invert_sigmas':False, 'max_image_seq_len':8192, 'max_shift':math.log(3), 'num_train_timesteps':1000, 'shift':1, 'shift_terminal':None, 'stochastic_sampling':False, 'time_shift_type':'exponential', 'use_beta_sigmas':False, 'use_dynamic_shifting':True, 'use_exponential_sigmas':False, 'use_karras_sigmas':False}) qwen.load_lora_weights('lightx2v/Qwen-Image-Lightning', weight_name='Qwen-Image-Lightning-4steps-V2.0.safetensors', adapter_name='lightning') qwen.set_adapters('lightning', adapter_weights=1) qwen(prompt='a beautiful girl', height=1280, width=720, num_inference_steps=4, true_cfg_scale=1).images[0].save('a.png') ``` WARNING:accelerate.big_modeling:Some parameters are on the meta device because they were offloaded to the cpu. {'text_encoder': 'cpu', 'vae': 0} where is the transformer ? NotImplementedError: Cannot copy out of meta tensor; no data! I want to ask how to make the above code work in kaggle. why 16G * 2 vram still not enough to run q4 quantization qwen-image? I want to take full advantage of 2 gpu. Do I need max_memory? full error logs: /usr/local/lib/python3.11/dist-packages/torch/utils/_contextlib.py in decorate_context(*args, **kwargs) 114 def decorate_context(*args, **kwargs): 115 with ctx_factory(): --> 116 return func(*args, **kwargs) 117 118 return decorate_context /usr/local/lib/python3.11/dist-packages/diffusers/pipelines/qwenimage/pipeline_qwenimage.py in __call__(self, prompt, negative_prompt, true_cfg_scale, height, width, num_inference_steps, sigmas, guidance_scale, num_images_per_prompt, generator, latents, prompt_embeds, prompt_embeds_mask, negative_prompt_embeds, negative_prompt_embeds_mask, output_type, return_dict, attention_kwargs, callback_on_step_end, callback_on_step_end_tensor_inputs, max_sequence_length) 566 ) 567 do_true_cfg = true_cfg_scale > 1 and has_neg_prompt --> 568 prompt_embeds, prompt_embeds_mask = self.encode_prompt( 569 prompt=prompt, 570 prompt_embeds=prompt_embeds, /usr/local/lib/python3.11/dist-packages/diffusers/pipelines/qwenimage/pipeline_qwenimage.py in encode_prompt(self, prompt, device, num_images_per_prompt, prompt_embeds, prompt_embeds_mask, max_sequence_length) 252 253 if prompt_embeds is None: --> 254 prompt_embeds, prompt_embeds_mask = self._get_qwen_prompt_embeds(prompt, device) 255 256 prompt_embeds = prompt_embeds[:, :max_sequence_length] /usr/local/lib/python3.11/dist-packages/diffusers/pipelines/qwenimage/pipeline_qwenimage.py in _get_qwen_prompt_embeds(self, prompt, device, dtype) 203 txt, max_length=self.tokenizer_max_length + drop_idx, padding=True, truncation=True, return_tensors="pt" 204 ).to(device) --> 205 encoder_hidden_states = self.text_encoder( 206 input_ids=txt_tokens.input_ids, 207 attention_mask=txt_tokens.attention_mask, /usr/local/lib/python3.11/dist-packages/torch/nn/modules/module.py in _wrapped_call_impl(self, *args, **kwargs) 1737 return self._compiled_call_impl(*args, **kwargs) # type: ignore[misc] 1738 else: -> 1739 return self._call_impl(*args, **kwargs) 1740 1741 # torchrec tests the code consistency with the following code /usr/local/lib/python3.11/dist-packages/torch/nn/modules/module.py in _call_impl(self, *args, **kwargs) 1748 or _global_backward_pre_hooks or _global_backward_hooks 1749 or _global_forward_hooks or _global_forward_pre_hooks): -> 1750 return forward_call(*args, **kwargs) 1751 1752 result = None /usr/local/lib/python3.11/dist-packages/accelerate/hooks.py in new_forward(module, *args, **kwargs) 173 output = module._old_forward(*args, **kwargs) 174 else: --> 175 output = module._old_forward(*args, **kwargs) 176 return module._hf_hook.post_forward(module, output) 177 /usr/local/lib/python3.11/dist-packages/transformers/utils/generic.py in wrapper(self, *args, **kwargs) 941 942 try: --> 943 output = func(self, *args, **kwargs) 944 if is_requested_to_return_tuple or (is_configured_to_return_tuple and is_top_level_module): 945
https://github.com/huggingface/diffusers/issues/12719
open
[]
2025-11-26T08:35:46Z
2025-11-26T09:15:54Z
null
chaowenguo
vllm-project/vllm
29,474
[P/D][Metrics] Consider combined/summed metrics (e.g. ttft and e2e_request_latency) for prefill and decode instances
### Your current environment <details> <summary>Env info snipped</summary> ``` Collecting environment information... uv is set ============================== System Info ============================== OS : Ubuntu 24.04.1 LTS (x86_64) GCC version : (Ubuntu 13.3.0-6ubuntu2~24.04) 13.3.0 Clang version : Could not collect CMake version : version 3.28.3 Libc version : glibc-2.39 ============================== PyTorch Info ============================== PyTorch version : 2.8.0+cu128 Is debug build : False CUDA used to build PyTorch : 12.8 ROCM used to build PyTorch : N/A ============================== Python Environment ============================== Python version : 3.12.3 (main, Aug 14 2025, 17:47:21) [GCC 13.3.0] (64-bit runtime) Python platform : Linux-5.15.0-152-generic-x86_64-with-glibc2.39 ============================== CUDA / GPU Info ============================== Is CUDA available : True CUDA runtime version : 12.8.93 CUDA_MODULE_LOADING set to : LAZY GPU models and configuration : GPU 0: NVIDIA H20 GPU 1: NVIDIA H20 GPU 2: NVIDIA H20 GPU 3: NVIDIA H20 GPU 4: NVIDIA H20 GPU 5: NVIDIA H20 GPU 6: NVIDIA H20 GPU 7: NVIDIA H20 Nvidia driver version : 570.172.08 cuDNN version : Probably one of the following: /usr/lib/x86_64-linux-gnu/libcudnn.so.9.8.0 /usr/lib/x86_64-linux-gnu/libcudnn_adv.so.9.8.0 /usr/lib/x86_64-linux-gnu/libcudnn_cnn.so.9.8.0 /usr/lib/x86_64-linux-gnu/libcudnn_engines_precompiled.so.9.8.0 /usr/lib/x86_64-linux-gnu/libcudnn_engines_runtime_compiled.so.9.8.0 /usr/lib/x86_64-linux-gnu/libcudnn_graph.so.9.8.0 /usr/lib/x86_64-linux-gnu/libcudnn_heuristic.so.9.8.0 /usr/lib/x86_64-linux-gnu/libcudnn_ops.so.9.8.0 HIP runtime version : N/A MIOpen runtime version : N/A Is XNNPACK available : True ============================== CPU Info ============================== Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 46 bits physical, 57 bits virtual Byte Order: Little Endian CPU(s): 128 On-line CPU(s) list: 0-127 Vendor ID: GenuineIntel BIOS Vendor ID: Intel(R) Corporation Model name: INTEL(R) XEON(R) PLATINUM 8562Y+ BIOS Model name: INTEL(R) XEON(R) PLATINUM 8562Y+ CPU @ 2.8GHz BIOS CPU family: 179 CPU family: 6 Model: 207 Thread(s) per core: 2 Core(s) per socket: 32 Socket(s): 2 Stepping: 2 CPU(s) scaling MHz: 73% CPU max MHz: 4100.0000 CPU min MHz: 800.0000 BogoMIPS: 5600.00 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf tsc_known_freq pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb cat_l3 cat_l2 cdp_l3 invpcid_single intel_ppin cdp_l2 ssbd mba ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid cqm rdt_a avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb intel_pt avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local split_lock_detect avx_vnni avx512_bf16 wbnoinvd dtherm ida arat pln pts hwp hwp_act_window hwp_epp hwp_pkg_req avx512vbmi umip pku ospke waitpkg avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg tme avx512_vpopcntdq la57 rdpid bus_lock_detect cldemote movdiri movdir64b enqcmd fsrm md_clear serialize tsxldtrk pconfig arch_lbr amx_bf16 avx512_fp16 amx_tile amx_int8 flush_l1d arch_capabilities Virtualization: VT-x L1d cache: 3 MiB (64 instances) L1i cache: 2 MiB (64 instances) L2 cache: 128 MiB (64 instances) L3 cache: 120 MiB (2 instances) NUMA node(s): 2 NUMA node0 CPU(s): 0-31,64-95 NUMA node1 CPU(s): 32-63,96-127 Vulnerability Gather data sampling: Not affected Vulnerability Indirect target s
https://github.com/vllm-project/vllm/issues/29474
open
[ "usage", "kv-connector" ]
2025-11-26T02:50:17Z
2025-11-26T08:31:18Z
1
mgw2168-1
vllm-project/vllm
29,472
[Installation]: how to Install vllm on dell promax gb10
### Your current environment I failed to install vllm on dell promax gb10 , mesages as followed nvcc --version nvcc: NVIDIA (R) Cuda compiler driver Copyright (c) 2005-2025 NVIDIA Corporation Built on Wed_Aug_20_01:57:39_PM_PDT_2025 Cuda compilation tools, release 13.0, V13.0.88 Build cuda_13.0.r13.0/compiler.36424714_0 pip install vllm Successfully installed torch-2.9.0 torchaudio-2.9.0 torchvision-0.24.0 vllm-0.11.2 ``` (py312) dell@promaxgb10-0843:~/test/vllm/Qwen$ vllm -V Traceback (most recent call last): File "/home/dell/miniconda3/envs/py312/bin/vllm", line 3, in <module> from vllm.entrypoints.cli.main import main File "/home/dell/miniconda3/envs/py312/lib/python3.12/site-packages/vllm/entrypoints/cli/__init__.py", line 3, in <module> from vllm.entrypoints.cli.benchmark.latency import BenchmarkLatencySubcommand File "/home/dell/miniconda3/envs/py312/lib/python3.12/site-packages/vllm/entrypoints/cli/benchmark/latency.py", line 5, in <module> from vllm.benchmarks.latency import add_cli_args, main File "/home/dell/miniconda3/envs/py312/lib/python3.12/site-packages/vllm/benchmarks/latency.py", line 17, in <module> from vllm.engine.arg_utils import EngineArgs File "/home/dell/miniconda3/envs/py312/lib/python3.12/site-packages/vllm/engine/arg_utils.py", line 35, in <module> from vllm.attention.backends.registry import AttentionBackendEnum File "/home/dell/miniconda3/envs/py312/lib/python3.12/site-packages/vllm/attention/__init__.py", line 4, in <module> from vllm.attention.backends.abstract import ( File "/home/dell/miniconda3/envs/py312/lib/python3.12/site-packages/vllm/attention/backends/abstract.py", line 9, in <module> from vllm.model_executor.layers.linear import ColumnParallelLinear File "/home/dell/miniconda3/envs/py312/lib/python3.12/site-packages/vllm/model_executor/__init__.py", line 4, in <module> from vllm.model_executor.parameter import BasevLLMParameter, PackedvLLMParameter File "/home/dell/miniconda3/envs/py312/lib/python3.12/site-packages/vllm/model_executor/parameter.py", line 11, in <module> from vllm.distributed import ( File "/home/dell/miniconda3/envs/py312/lib/python3.12/site-packages/vllm/distributed/__init__.py", line 4, in <module> from .communication_op import * File "/home/dell/miniconda3/envs/py312/lib/python3.12/site-packages/vllm/distributed/communication_op.py", line 9, in <module> from .parallel_state import get_tp_group File "/home/dell/miniconda3/envs/py312/lib/python3.12/site-packages/vllm/distributed/parallel_state.py", line 250, in <module> direct_register_custom_op( File "/home/dell/miniconda3/envs/py312/lib/python3.12/site-packages/vllm/utils/torch_utils.py", line 640, in direct_register_custom_op from vllm.platforms import current_platform File "/home/dell/miniconda3/envs/py312/lib/python3.12/site-packages/vllm/platforms/__init__.py", line 257, in __getattr__ _current_platform = resolve_obj_by_qualname(platform_cls_qualname)() ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/dell/miniconda3/envs/py312/lib/python3.12/site-packages/vllm/utils/import_utils.py", line 89, in resolve_obj_by_qualname module = importlib.import_module(module_name) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/dell/miniconda3/envs/py312/lib/python3.12/importlib/__init__.py", line 90, in import_module return _bootstrap._gcd_import(name[level:], package, level) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/dell/miniconda3/envs/py312/lib/python3.12/site-packages/vllm/platforms/cuda.py", line 16, in <module> import vllm._C # noqa ^^^^^^^^^^^^^^ ImportError: libtorch_cuda.so: cannot open shared object file: No such file or directory ``` ### How you are installing vllm ```sh pip install vllm ``` ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29472
open
[ "installation" ]
2025-11-26T02:41:18Z
2026-01-01T12:28:29Z
2
goactiongo
vllm-project/vllm
29,436
[Bug]: vLLM Serve with LMCache enabled produces wrong output for GPT-OSS-20B
### Your current environment <details> <summary>The output of <code>python collect_env.py</code></summary> ```text Your output of `python collect_env.py` here ``` </details> ### 🐛 Describe the bug vLLM serve command with LMCache enabled produces wrong output with GPT OSS 20B for subsequent invocations with the same prompt Steps to reproduce: Command to start the server: ``` LMCACHE_CONFIG_FILE=lmcache_cpu.yaml vllm serve openai/gpt-oss-20b --port 8000 --kv-transfer-config '{"kv_connector":"LMCacheConnectorV1", "kv_role":"kv_both"}' ``` Invocation: ``` curl 127.0.0.1:8000/v1/chat/completions -H "Content-Type: application/json" -d '{"model": "openai/gpt-oss-20b", "messages": [ {"role": "user", "content": "What is Amazon SageMaker?"}]}' ``` First invocation: ``` { "id":"chatcmpl-951ca7178b1e4226b0343cb070033487", "object":"chat.completion", "created":1764098087, "model":"openai/gpt-oss-20b", "choices":[ {"index":0,"message":{"role":"assistant","content":"**Amazon SageMaker** is Amazon Web Services’ fully‑managed platform that lets you build, train, tune, and deploy machine‑learning models fast—without managing the underlying infrastructure.\n\nKey capabilities\n\n| Feature | What it does |\n|--------|--------------|\n| **SageMaker Studio** | A web‑based IDE that bundles notebooks, visual debugging, model monitoring, and collaboration tools. |\n| **Built‑in algorithms & frameworks** | Pre‑packaged models (XGBoost, Linear Learner, etc.) and support for your own TensorFlow, PyTorch, MXNet, Scikit‑learn, R, etc. |\n| **Auto‑ML & automated model tuning** | SageMaker Autopilot automatically searches model architectures and hyper‑parameters. |\n| **Managed training** | Spot, distributed, and GPU training jobs that scale to the required compute. |\n| **Model deployment** | One‑click production endpoints, batch transform, edge inference (SageMaker Edge), and real‑time or asynchronous inference. |\n| **Inference pipelines** | Compose multiple models or processing steps into a single pipeline. |\n| **Model monitoring & A/B testing** | Continuous evaluation of drift, predictions, and performance metrics. |\n| **Security & compliance** | VPC, IAM, KMS encryption, private cataloging, and audit trails. |\n\nIn short, SageMaker removes the operational burden of ML—so teams can focus on data science and business value rather than servers, networking, and scaling.","refusal":null,"annotations":null,"audio":null,"function_call":null,"tool_calls":[],"reasoning":"User asks \"What is Amazon SageMaker?\" Short answer. Provide description: fully managed ML service, environment to build, train, deploy models, etc. Should be succinct.","reasoning_content":"User asks \"What is Amazon SageMaker?\" Short answer. Provide description: fully managed ML service, environment to build, train, deploy models, etc. Should be succinct."},"logprobs":null,"finish_reason":"stop","stop_reason":null,"token_ids":null}],"service_tier":null,"system_fingerprint":null,"usage":{"prompt_tokens":75,"total_tokens":426,"completion_tokens":351,"prompt_tokens_details":null},"prompt_logprobs":null,"prompt_token_ids":null,"kv_transfer_params":null} ``` Second invocation: ``` { "id": "chatcmpl-4ebc19fc5c2a41a7bebc01ea8d1c98b1", "object": "chat.completion", "created": 1764098160, "model": "openai/gpt-oss-20b", "choices": [ { "index": 0, "message": { "role": "assistant", "content": "Sure! Here’s a basic guide to get you started with writing a cool, informative yet accessible article on **\"The Fascinating World of Quantum Computing\"** for a general audience. Feel free to adapt the structure, tone, or content to match your style and publication’s guidelines.\n\n---\n\n## 1. Hook & Context (≈150–200 words)\n\n- **Start with a vivid anecdote, surprising fact, or a relatable analogy** that introduces the “wow” moment in quantum computing.\n - *Example:* “Imagine a coin that, instead of being heads or tails, can be both at the same time… until you look at it.” \n- **Briefly state why this topic matters** to everyday life: faster drug discovery, better encryption, breakthrough materials, etc.\n\n> **Tell readers what they’ll learn**: a quick glimpse of quantum fundamentals, why it’s different from classic bits, and how it could reshape technologies.\n\n---\n\n## 2. What’s a Quantum Computer? (≈300 words)\n\n| Section | Content | Quick Tips |\n|---------|---------|------------|\n| **2.1 “Bits” vs. “Qubits”** | • Classical bits (“0” or “1”).<br>• Qubits: superposition (both 0 & 1) & entanglement. | Use visual metaphors: a spinning top (superposition) and two dancers always in sync (entanglement). |\n| **2.2 Basic Operations** | • Quantum gates (Pauli X, H, CNOT).<br>• The role of interference. | A tiny “reversible” logic of the quantum “if‑then” that flips outcomes. |\n| **2.3 Measuring As a Collapses** | • Outcome collapse on measurement.<br>• Probabilities & expectation values. | Compare to a gamble: you only learn the re
https://github.com/vllm-project/vllm/issues/29436
open
[ "bug" ]
2025-11-25T19:27:24Z
2025-11-25T19:27:24Z
0
ksuma2109
vllm-project/vllm
29,409
[Usage]: Custom Logits Processors V1 how to get tokenizer into processor
### Problem with tokenizer For the second day now, I've been unable to figure out how to get a tokenizer inside a custom processor. I used the processor from the documentation as an example. I examined each object through debug, but couldn't find where to extract the tokenizer. In v0, this was done simply at the request level, by passing an argument to the object. How to pass a tokenizer to the processor? ```python import torch from vllm.config import VllmConfig from vllm.sampling_params import SamplingParams from vllm.v1.sample.logits_processor import (BatchUpdate, LogitsProcessor, MoveDirectionality) class DummyLogitsProcessor(LogitsProcessor): """Fake logit processor to support unit testing and examples""" @classmethod def validate_params(cls, params: SamplingParams): target_token: int | None = params.extra_args and params.extra_args.get( "target_token" ) if target_token is not None and not isinstance(target_token, int): raise ValueError(f"target_token value {target_token} is not int") def __init__(self, vllm_config: "VllmConfig", device: torch.device, is_pin_memory: bool): self.req_info: dict[int, int] = {} def is_argmax_invariant(self) -> bool: """Never impacts greedy sampling""" return False def update_state(self, batch_update: BatchUpdate | None): if not batch_update: return # Process added requests. for index, params, _, _ in batch_update.added: assert params is not None self.validate_params(params) if params.extra_args and (target_token := params.extra_args.get("target_token")): self.req_info[index] = target_token else: self.req_info.pop(index, None) if self.req_info: # Process removed requests. for index in batch_update.removed: self.req_info.pop(index, None) # Process moved requests, unidirectional move (a->b) and swap # (a<->b) for adx, bdx, direct in batch_update.moved: a_val = self.req_info.pop(adx, None) b_val = self.req_info.pop(bdx, None) if a_val is not None: self.req_info[bdx] = a_val if direct == MoveDirectionality.SWAP and b_val is not None: self.req_info[adx] = b_val def apply(self, logits: torch.Tensor) -> torch.Tensor: if not self.req_info: return logits # Save target values before modification cols = torch.tensor( list(self.req_info.values()), dtype=torch.long, device=logits.device ) rows = torch.tensor( list(self.req_info.keys()), dtype=torch.long, device=logits.device ) values_to_keep = logits[rows, cols].clone() # Mask all but target tokens logits[rows] = float('-inf') logits[rows, cols] = values_to_keep return logits ``` ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29409
closed
[ "usage" ]
2025-11-25T13:24:17Z
2025-12-02T10:33:18Z
6
cvadim130
vllm-project/vllm
29,389
[Bug]: race condition in shm_broadcast.py
### Your current environment <details> <summary>The output of <code>python collect_env.py</code></summary> ```text Your output of `python collect_env.py` here ``` </details> ### 🐛 Describe the bug # Problem `ShmRingBuffer` is a lock-free queue, the implementation of which https://github.com/vllm-project/vllm/blob/12c007e288bf5c0ae3bd438036fbafbad88e706b/vllm/distributed/device_communicators/shm_broadcast.py#L98-L153 relies on the fact that when a flag is written to, signalling a valid state, the associated data is also in a valid state. To illustrate the point, consider the program ```python shm = shared_memory.SharedMemory(..., size=128) # set shm to 0 # process 1 shm[0] = 1 shm[64] = 1 # process 2 while shm[64] != 1: pass print(shm[0]) ``` `ShmRingBuffer` requires that `print(shm[0])` always prints `1`. **There is no guarantee this is true**. For this to be true, 1. The Python language/implementation must provide a memory model, which it doesn't. Loosely speaking, a memory model is a set of guarantees on how source code maps to hardware instructions. 2. Even if we assume the source code maps "as intended" to hardware instructions, the hardware must ensure that process 2 must observe the writes to `shm[0]` and `shm[64]` in the same order as process 1. An example of 2 breaking down is given in [`race_condition.cpp`](https://gist.github.com/nvjullin/cc52386e291fe41218b54406ece962a0). On an ARM CPU, ```bash $ g++ -std=c++17 race_condition.cpp $ ./a.out number of violations: 5 # ... ``` Unfortunately, I don't know how to demonstrate the same race condition in Python. # What it means `ShmRingBuffer` can have corrupted memory and crashes vLLM sporadically. Such a crash would be near impossible to reproduce and debug. # Solutions In order of recommendation: 1. Remove `ShmRingBuffer` and always use the fallback `self.local_socket.send(serialized_obj)`. This is the simplest. 2. Use a well-tested lock-free queue implementation and don't write our own. Lock-free programming is notoriously difficult to write correctly, requires expertise to understand and is overall a maintenence nightmare. 3. Write it in C++ with proper atomics that guarantees the ordering of writes. The implementation should document extensively the proof of its correctness across different architectures. Python provides no tools for lock-free programming, making it impossible to write. CC @youkaichao @nvpohanh ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29389
open
[ "bug" ]
2025-11-25T09:25:52Z
2025-11-25T09:25:52Z
0
nvjullin
vllm-project/vllm
29,382
[Doc]: Expert Parallel Deployment says "Tensor parallel size (always 1 for now)" is confusing
### 📚 The doc issue On page https://docs.vllm.ai/en/latest/serving/expert_parallel_deployment/#single-node-deployment it says Tensor parallel size can only be 1 but didn't mention the behavior of Attention Layers On page https://docs.vllm.ai/en/latest/serving/data_parallel_deployment/ it says The expert layers will by default form a (DP x TP) sized tensor parallel group. To enable expert parallelism, include the --enable-expert-parallel CLI arg (on all nodes in the multi-node case). which is rather confusing. ### Suggest a potential alternative/fix Point out the correct behavior of MoE models when TP, EP are both set. ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29382
closed
[ "documentation" ]
2025-11-25T07:54:42Z
2025-12-13T17:38:01Z
0
xeonliu
huggingface/transformers
42,375
SAM3 single image inference with multiple text prompt
Hi I'm trying to run inference on a single image, aiming to get the bbox of objects from several different categories (e.g. "a person" and "a car"). the only example i found for prompting with multiple categories is in the "Batched Inference with Text Prompts" example, but then i need to unnecessarily duplicate my image as the # of categories. is there a different more efficient way of achieving this? p.s when i try prompting with a list of several categories and a single image i get an error.
https://github.com/huggingface/transformers/issues/42375
open
[]
2025-11-25T06:20:09Z
2026-01-05T16:16:01Z
9
iariav
huggingface/trl
4,569
[doc issue] doc on "GRPO with replay buffer" buggy
### Reproduction The code example in [doc for "GRPO with replay buffer"](https://huggingface.co/docs/trl/main/en/experimental#grpo-with-replay-buffer) is kind of buggy. - It imports `GRPOWithReplayBufferTrainer` but never used. - It uses `GRPOWithReplayBufferConfig` but never imported - The code is apparently not executable. Below is the code example given in the doc: ```python from trl.experimental.grpo_with_replay_buffer import GRPOWithReplayBufferTrainer from datasets import load_dataset dataset = load_dataset("trl-internal-testing/zen", "standard_prompt_only", split="train") # Guarantee that some rewards have 0 std def custom_reward_func(completions, **kwargs): if torch.rand(1).item() < 0.25: return [0] * len(completions) # simulate some None rewards else: return torch.rand(len(completions)).tolist() training_args = GRPOWithReplayBufferConfig( output_dir=self.tmp_dir, learning_rate=1e-4, per_device_train_batch_size=4, num_generations=4, max_completion_length=8, replay_buffer_size=8, report_to="none", ) trainer = GRPOTrainer( model="trl-internal-testing/tiny-Qwen2ForCausalLM-2.5", reward_funcs=[custom_reward_func], args=training_args, train_dataset=dataset, ) previous_trainable_params = {n: param.clone() for n, param in trainer.model.named_parameters()} trainer.train() ``` ### System Info NA ### Checklist - [x] I have checked that my issue isn't already filed (see [open issues](https://github.com/huggingface/trl/issues?q=is%3Aissue)) - [x] I have included my system information - [x] Any code provided is minimal, complete, and reproducible ([more on MREs](https://docs.github.com/en/get-started/writing-on-github/working-with-advanced-formatting/creating-and-highlighting-code-blocks)) - [x] Any code provided is properly formatted in code blocks, (no screenshot, [more on code blocks](https://docs.github.com/en/get-started/writing-on-github/working-with-advanced-formatting/creating-and-highlighting-code-blocks)) - [x] Any traceback provided is complete
https://github.com/huggingface/trl/issues/4569
closed
[ "🐛 bug", "📚 documentation", "🏋 GRPO" ]
2025-11-25T01:30:28Z
2025-11-25T21:28:00Z
2
DNXie
vllm-project/vllm
29,306
[Usage]: dots.llm.inst is not running due to a type error
### Your current environment I'm trying to run dots llm on 4xH100 ``` vllm serve \ --uvicorn-log-level=info \ rednote-hilab/dots.llm1.inst \ --dtype auto \ --api-key xxx \ --host 0.0.0.0 \ --port 8000 \ --tensor-parallel-size 4 --ipc=host \ --trust-remote-code ``` It failed to run, I got the following crash: ```text (EngineCore_DP0 pid=10684) ERROR 11-24 09:41:25 [v1/executor/multiproc_executor.py:230] Worker proc VllmWorker-1 died unexpectedly, shutting down executor. (EngineCore_DP0 pid=10684) Process EngineCore_DP0: (EngineCore_DP0 pid=10684) Traceback (most recent call last): (EngineCore_DP0 pid=10684) File "/usr/lib/python3.12/multiprocessing/process.py", line 314, in _bootstrap (EngineCore_DP0 pid=10684) self.run() (EngineCore_DP0 pid=10684) File "/usr/lib/python3.12/multiprocessing/process.py", line 108, in run (EngineCore_DP0 pid=10684) self._target(*self._args, **self._kwargs) (EngineCore_DP0 pid=10684) File "/home/ubuntu/venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 846, in run_engine_core (EngineCore_DP0 pid=10684) raise e (EngineCore_DP0 pid=10684) File "/home/ubuntu/venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 833, in run_engine_core (EngineCore_DP0 pid=10684) engine_core = EngineCoreProc(*args, **kwargs) (EngineCore_DP0 pid=10684) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ (EngineCore_DP0 pid=10684) File "/home/ubuntu/venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 606, in __init__ (EngineCore_DP0 pid=10684) super().__init__( (EngineCore_DP0 pid=10684) File "/home/ubuntu/venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 109, in __init__ (EngineCore_DP0 pid=10684) num_gpu_blocks, num_cpu_blocks, kv_cache_config = self._initialize_kv_caches( (EngineCore_DP0 pid=10684) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ (EngineCore_DP0 pid=10684) File "/home/ubuntu/venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 231, in _initialize_kv_caches (EngineCore_DP0 pid=10684) available_gpu_memory = self.model_executor.determine_available_memory() (EngineCore_DP0 pid=10684) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ (EngineCore_DP0 pid=10684) File "/home/ubuntu/venv/lib/python3.12/site-packages/vllm/v1/executor/abstract.py", line 126, in determine_available_memory (EngineCore_DP0 pid=10684) return self.collective_rpc("determine_available_memory") (EngineCore_DP0 pid=10684) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ (EngineCore_DP0 pid=10684) File "/home/ubuntu/venv/lib/python3.12/site-packages/vllm/v1/executor/multiproc_executor.py", line 358, in collective_rpc (EngineCore_DP0 pid=10684) return aggregate(get_response()) (EngineCore_DP0 pid=10684) ^^^^^^^^^^^^^^ (EngineCore_DP0 pid=10684) File "/home/ubuntu/venv/lib/python3.12/site-packages/vllm/v1/executor/multiproc_executor.py", line 341, in get_response (EngineCore_DP0 pid=10684) raise RuntimeError( (EngineCore_DP0 pid=10684) RuntimeError: Worker failed with error 'TypeError: can't multiply sequence by non-int of type 'float' (EngineCore_DP0 pid=11385) ERROR 11-24 09:45:27 [v1/engine/core.py:842] EngineCore failed to start. (EngineCore_DP0 pid=11385) ERROR 11-24 09:45:27 [v1/engine/core.py:842] Traceback (most recent call last): (EngineCore_DP0 pid=11385) ERROR 11-24 09:45:27 [v1/engine/core.py:842] File "/home/ubuntu/venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 833, in run_engine_core (EngineCore_DP0 pid=11385) ERROR 11-24 09:45:27 [v1/engine/core.py:842] engine_core = EngineCoreProc(*args, **kwargs) (EngineCore_DP0 pid=11385) ERROR 11-24 09:45:27 [v1/engine/core.py:842] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ (EngineCore_DP0 pid=11385) ERROR 11-24 09:45:27 [v1/engine/core.py:842] File "/home/ubuntu/venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 606, in __init__ (EngineCore_DP0 pid=11385) ERROR 11-24 09:45:27 [v1/engine/core.py:842] super().__init__( (EngineCore_DP0 pid=11385) ERROR 11-24 09:45:27 [v1/engine/core.py:842] File "/home/ubuntu/venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 109, in __init__ (EngineCore_DP0 pid=11385) ERROR 11-24 09:45:27 [v1/engine/core.py:842] num_gpu_blocks, num_cpu_blocks, kv_cache_config = self._initialize_kv_caches( (EngineCore_DP0 pid=11385) ERROR 11-24 09:45:27 [v1/engine/core.py:842] ^^^^^^^^^^^^^^^^^^^^^^^^^^^ (EngineCore_DP0 pid=11385) ERROR 11-24 09:45:27 [v1/engine/core.py:842] File "/home/ubuntu/venv/lib/python3.12/site-packages/vllm/v1/engine/core.py", line 231, in _initialize_kv_caches (EngineCore_DP0 pid=11385) ERROR 11-24 09:45:27 [v1/engine/core.py:842] available_gpu_memory = self.model_executor.determine_available_memory() (EngineCore_DP0 pid=11385) ERROR 11-24 09:45:27 [v1/engine/core.py:842]
https://github.com/vllm-project/vllm/issues/29306
closed
[ "usage" ]
2025-11-24T09:48:08Z
2025-11-28T23:25:27Z
1
rain-1
huggingface/transformers
42,353
SAM3 point mode is not supported yet?
In [SAM3 official example](https://github.com/facebookresearch/sam3/blob/main/examples/sam3_for_sam1_task_example.ipynb ), they also support point mode. But it seems that transforms has not supported yet?
https://github.com/huggingface/transformers/issues/42353
closed
[]
2025-11-24T07:16:52Z
2025-11-26T15:16:25Z
1
haofanwang
vllm-project/vllm
29,297
[Bug]: What should the image embedding input be like? I have tested with multiple cases but it all fails
### Your current environment ```text ============================== System Info ============================== OS : Red Hat Enterprise Linux release 8.10 (Ootpa) (x86_64) GCC version : (GCC) 8.5.0 20210514 (Red Hat 8.5.0-26) Clang version : Could not collect CMake version : Could not collect Libc version : glibc-2.28 ============================== PyTorch Info ============================== PyTorch version : 2.9.0+cu128 Is debug build : False CUDA used to build PyTorch : 12.8 ROCM used to build PyTorch : N/A ============================== Python Environment ============================== Python version : 3.10.19 (main, Oct 21 2025, 16:43:05) [GCC 11.2.0] (64-bit runtime) Python platform : Linux-4.18.0-553.50.1.el8_10.x86_64-x86_64-with-glibc2.28 ============================== CUDA / GPU Info ============================== Is CUDA available : True CUDA runtime version : Could not collect CUDA_MODULE_LOADING set to : GPU models and configuration : GPU 0: NVIDIA A100-SXM4-40GB GPU 1: NVIDIA A100-SXM4-40GB Nvidia driver version : 575.51.03 cuDNN version : Could not collect HIP runtime version : N/A MIOpen runtime version : N/A Is XNNPACK available : True ============================== CPU Info ============================== Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 128 On-line CPU(s) list: 0-127 Thread(s) per core: 1 Core(s) per socket: 64 Socket(s): 2 NUMA node(s): 8 Vendor ID: AuthenticAMD CPU family: 23 Model: 49 Model name: AMD EPYC 7742 64-Core Processor Stepping: 0 CPU MHz: 2250.000 CPU max MHz: 2250.0000 CPU min MHz: 1500.0000 BogoMIPS: 4491.72 Virtualization: AMD-V L1d cache: 32K L1i cache: 32K L2 cache: 512K L3 cache: 16384K NUMA node0 CPU(s): 0-15 NUMA node1 CPU(s): 16-31 NUMA node2 CPU(s): 32-47 NUMA node3 CPU(s): 48-63 NUMA node4 CPU(s): 64-79 NUMA node5 CPU(s): 80-95 NUMA node6 CPU(s): 96-111 NUMA node7 CPU(s): 112-127 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid aperfmperf pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_llc mwaitx cpb cat_l3 cdp_l3 hw_pstate ssbd mba ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr wbnoinvd arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif v_spec_ctrl umip rdpid overflow_recov succor smca sme sev sev_es ============================== Versions of relevant libraries ============================== [pip3] flashinfer-python==0.5.2 [pip3] numpy==2.2.6 [pip3] nvidia-cublas-cu12==12.8.4.1 [pip3] nvidia-cuda-cupti-cu12==12.8.90 [pip3] nvidia-cuda-nvrtc-cu12==12.8.93 [pip3] nvidia-cuda-runtime-cu12==12.8.90 [pip3] nvidia-cudnn-cu12==9.10.2.21 [pip3] nvidia-cudnn-frontend==1.16.0 [pip3] nvidia-cufft-cu12==11.3.3.83 [pip3] nvidia-cufile-cu12==1.13.1.3 [pip3] nvidia-curand-cu12==10.3.9.90 [pip3] nvidia-cusolver-cu12==11.7.3.90 [pip3] nvidia-cusparse-cu12==12.5.8.93 [pip3] nvidia-cusparselt-cu12==0.7.1 [pip3] nvidia-cutlass-dsl==4.3.0 [pip3] nvidia-ml-py==13.580.82 [pip3] nvidia-nccl-cu12==2.27.5 [pip3] nvidia-nvjitlink-cu12==12.8.93 [pip3] nvidia-nvshmem-cu12==3.3.20 [pip3] nvidia-nvtx-cu12==12.8.90 [pip3] pyzmq==27.1.0 [pip3] torch==2.9.0 [pip3] torchaudio==2.9.0 [pip3] torchvision==0.24.0 [pip3] transformers==4.57.1 [pip3] triton==3.5.0 [conda] flashinfer-python 0.5.2 pypi_0 pypi [conda] numpy 2.2.6 pypi_0 pypi [conda] nvidia-cublas-cu12 12.8.4.1 pypi_0 pypi [conda] nvidia-cuda-cupti-cu12 12.8.90 pypi_0 pypi [conda] nvidia-cuda-nvrtc-cu12 12.8.93 pypi_0 pypi [conda] nvidia-cuda-runtime-cu12 12.8.90 pypi_0 pypi [conda] nvidia-cudnn-cu12 9.10.2.21 pypi_0 pypi [conda] nvidia-cudn
https://github.com/vllm-project/vllm/issues/29297
closed
[ "usage" ]
2025-11-24T06:02:09Z
2025-11-26T13:00:17Z
2
DamonZhao-sfu
vllm-project/vllm
29,294
[CPU Backend] [Doc]: Update Installation Docs for Arm CPUs
### 📚 The doc issue This page https://docs.vllm.ai/en/stable/getting_started/installation/cpu/#arm-aarch64 is very out-dated. We now release Arm CPU wheels and images thanks to #26931 and #27331 We need to update that page to reflect that :) ### Suggest a potential alternative/fix _No response_ ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29294
closed
[ "documentation", "cpu" ]
2025-11-24T05:33:46Z
2025-12-15T19:46:26Z
5
fadara01
vllm-project/vllm
29,286
[Performance]: cache system prompt token ids
### Proposal to improve performance As system prompt can be very long now, tokenize the system prompt can be slow. Using H20, tokenize 5000 tokens cost about 10ms as below: ![Image](https://github.com/user-attachments/assets/e1b0dafa-6514-47e6-8531-db8eaea32cc7) System prompts are usually fixed and reusable, so cache the system prompt can be profitable. Specificly: 1. In **apply_hf_chat_template** method we can separate the system prompt from other prompts, we can use condition **cache_system_prompt = truncate_prompt_tokens is None and not tokenize and len(conversation) > 1 and conversation[0].get("role") == "system"** to judge when we should separate the system prompt. 2. In **_normalize_prompt_text_to_input** method we judge that whether system prompt is in the dict ({system prompt: token ids}) that we can reuse, then concat system prompt token ids and prompt token ids as the final input_ids. I am willing to contribute to this opt and looking forward to your suggestions! ### Report of performance regression The above cost can be profitable. ### Misc discussion on performance _No response_ ### Your current environment (if you think it is necessary) ```text The output of `python collect_env.py` ``` ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29286
open
[ "performance" ]
2025-11-24T01:55:32Z
2025-11-28T08:57:06Z
2
Eviannn
vllm-project/vllm
29,281
[Usage]: Removing last generated token from output and kv cache
### Your current environment ```text vLLM 0.11.2 ``` ### How would you like to use vllm Hey guys, i am currently working on a research project where i load a moe-like model and i want to do routing based on the sequence state. The goal is to let expert 0 generate until it reaches the eos token, then remove the eos token and finish generation with expert 1 until the eos token is hit a second time. I want to do this to use different strengths of both models. My current approach is to modify GPUModelRunner and Scheduler to remove the eos token from output, reduce num_computed_tokens by 1 and compute a static routing tensor based on the sequence state which i pass as additional model input, to route to expert 0 or 1. Now i am having some issues with unexpected output, especially with tensor_parallelism>1 on multiple gpus. I was wondering if there already is a reliable solution to remove the last generated token from output and kv cache, so that the computation leading to eos does not interfere with the second expert. Or maybe there is even a better way to do this? Thank you! ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29281
closed
[ "usage" ]
2025-11-23T22:39:16Z
2025-11-26T09:33:53Z
0
josefdra
vllm-project/vllm
29,277
[Usage]: Creating and accessing per request arguments inside vLLM model
### Your current environment ```text The output of `python collect_env.py` ``` ### How would you like to use vllm I want to implement token compression techniques on the output embeddings of Qwen-2.5VL which would occur dynamically as the number of requests change. Is there anyway to implement this in vLLM? I see that SamplingParams seem to be the only way to use per request custom arguments but I don’t believe it can be accessed within the model code directly? ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29277
open
[ "usage" ]
2025-11-23T21:59:31Z
2025-11-23T21:59:31Z
0
minlu21
huggingface/transformers
42,344
How to fine-tune SAM 3D models?
### Model description The recently released SAM 3D work is truly remarkable. Do you plan to integrate it into Transformers and enable fine-tuning? https://huggingface.co/facebook/sam-3d-objects ### Open source status - [x] The model implementation is available - [x] The model weights are available ### Provide useful links for the implementation _No response_
https://github.com/huggingface/transformers/issues/42344
open
[ "New model" ]
2025-11-23T17:40:57Z
2025-11-23T17:40:57Z
null
bruno686
vllm-project/vllm
29,264
[Usage]: Monkey Patching SamplingParams
### Your current environment ```text Collecting environment information... ============================== System Info ============================== OS : Ubuntu 24.04.3 LTS (x86_64) GCC version : (Ubuntu 13.3.0-6ubuntu2~24.04) 13.3.0 Clang version : Could not collect CMake version : version 3.28.3 Libc version : glibc-2.39 ============================== PyTorch Info ============================== PyTorch version : 2.9.0+cu128 Is debug build : False CUDA used to build PyTorch : 12.8 ROCM used to build PyTorch : N/A ============================== Python Environment ============================== Python version : 3.13.5 | packaged by conda-forge | (main, Jun 16 2025, 08:27:50) [GCC 13.3.0] (64-bit runtime) Python platform : Linux-6.8.0-87-generic-x86_64-with-glibc2.39 ============================== CUDA / GPU Info ============================== Is CUDA available : True CUDA runtime version : 12.8.93 CUDA_MODULE_LOADING set to : GPU models and configuration : GPU 0: NVIDIA B200 GPU 1: NVIDIA B200 GPU 2: NVIDIA B200 GPU 3: NVIDIA B200 GPU 4: NVIDIA B200 GPU 5: NVIDIA B200 GPU 6: NVIDIA B200 GPU 7: NVIDIA B200 Nvidia driver version : 570.195.03 cuDNN version : Could not collect HIP runtime version : N/A MIOpen runtime version : N/A Is XNNPACK available : True ============================== CPU Info ============================== Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 52 bits physical, 57 bits virtual Byte Order: Little Endian CPU(s): 224 On-line CPU(s) list: 0-223 Vendor ID: GenuineIntel Model name: INTEL(R) XEON(R) PLATINUM 8570 CPU family: 6 Model: 207 Thread(s) per core: 2 Core(s) per socket: 56 Socket(s): 2 Stepping: 2 CPU(s) scaling MHz: 31% CPU max MHz: 4000.0000 CPU min MHz: 800.0000 BogoMIPS: 4200.00 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf tsc_known_freq pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb cat_l3 cat_l2 cdp_l3 intel_ppin cdp_l2 ssbd mba ibrs ibpb stibp ibrs_enhanced tpr_shadow flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid cqm rdt_a avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb intel_pt avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local split_lock_detect user_shstk avx_vnni avx512_bf16 wbnoinvd dtherm ida arat pln pts hwp hwp_act_window hwp_epp hwp_pkg_req vnmi avx512vbmi umip pku ospke waitpkg avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg tme avx512_vpopcntdq la57 rdpid bus_lock_detect cldemote movdiri movdir64b enqcmd fsrm md_clear serialize tsxldtrk pconfig arch_lbr ibt amx_bf16 avx512_fp16 amx_tile amx_int8 flush_l1d arch_capabilities ibpb_exit_to_user Virtualization: VT-x L1d cache: 5.3 MiB (112 instances) L1i cache: 3.5 MiB (112 instances) L2 cache: 224 MiB (112 instances) L3 cache: 600 MiB (2 instances) NUMA node(s): 2 NUMA node0 CPU(s): 0-55,112-167 NUMA node1 CPU(s): 56-111,168-223 Vulnerability Gather data sampling: Not affected Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Not affected Vulnerability Reg file data sampling: Not affected Vulnerability Retbleed: Not affected Vulnerability Spec rstack overflow: Not affected Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; Enhanced / Automatic IBRS; IBPB conditional; RSB filling; PBRSB-eIBRS SW sequence; BHI
https://github.com/vllm-project/vllm/issues/29264
closed
[ "usage" ]
2025-11-23T11:45:54Z
2025-11-24T13:03:50Z
2
josefdra
vllm-project/vllm
29,263
[Feature]: Enable flash attention (and/or FlashMLA) for AMD GPUs
### 🚀 The feature, motivation and pitch In [this page from flash-attention](https://github.com/Dao-AILab/flash-attention?tab=readme-ov-file#amd-rocm-support), I checked that the upstream `flash-attention` currently has composable_kernel (for newer AMD GPUs) and WIP triton (for older RNDA GPUs, etc.) implementations. As well as [flash MLA](https://github.com/deepseek-ai/FlashMLA?tab=readme-ov-file#amd-instinct). Is it possible to enable `vllm.vllm_flash_attn._vllm_fa2_C` and more modules for AMD GPUs? ### Alternatives _No response_ ### Additional context _No response_ ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29263
closed
[ "feature request", "rocm" ]
2025-11-23T11:28:47Z
2025-12-05T01:54:08Z
4
Inokinoki
vllm-project/vllm
29,245
[Usage]: 启动 qwen3 vl 超级超级超级慢,sglang 启动很快,可能的原因是什么?
### Your current environment 连执行 python collect_env.py 都很慢,环境是直接 uv 安装的 ```text Collecting environment information... ============================== System Info ============================== OS : Ubuntu 24.04.2 LTS (x86_64) GCC version : (Ubuntu 13.3.0-6ubuntu2~24.04) 13.3.0 Clang version : Could not collect CMake version : version 4.1.2 Libc version : glibc-2.39 ============================== PyTorch Info ============================== PyTorch version : 2.9.0+cu128 Is debug build : False CUDA used to build PyTorch : 12.8 ROCM used to build PyTorch : N/A ============================== Python Environment ============================== Python version : 3.12.3 (main, Jun 18 2025, 17:59:45) [GCC 13.3.0] (64-bit runtime) Python platform : Linux-5.10.134-19.100.al8.x86_64-x86_64-with-glibc2.39 ============================== CUDA / GPU Info ============================== Is CUDA available : True CUDA runtime version : 12.9.86 CUDA_MODULE_LOADING set to : GPU models and configuration : GPU 0: NVIDIA L20Y GPU 1: NVIDIA L20Y GPU 2: NVIDIA L20Y GPU 3: NVIDIA L20Y GPU 4: NVIDIA L20Y GPU 5: NVIDIA L20Y GPU 6: NVIDIA L20Y GPU 7: NVIDIA L20Y Nvidia driver version : 570.148.08 cuDNN version : Probably one of the following: /usr/lib/x86_64-linux-gnu/libcudnn.so.9.10.2 /usr/lib/x86_64-linux-gnu/libcudnn_adv.so.9.10.2 /usr/lib/x86_64-linux-gnu/libcudnn_cnn.so.9.10.2 /usr/lib/x86_64-linux-gnu/libcudnn_engines_precompiled.so.9.10.2 /usr/lib/x86_64-linux-gnu/libcudnn_engines_runtime_compiled.so.9.10.2 /usr/lib/x86_64-linux-gnu/libcudnn_graph.so.9.10.2 /usr/lib/x86_64-linux-gnu/libcudnn_heuristic.so.9.10.2 /usr/lib/x86_64-linux-gnu/libcudnn_ops.so.9.10.2 HIP runtime version : N/A MIOpen runtime version : N/A Is XNNPACK available : True ============================== CPU Info ============================== Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 46 bits physical, 57 bits virtual Byte Order: Little Endian CPU(s): 192 On-line CPU(s) list: 0-191 Vendor ID: GenuineIntel Model name: Intel(R) Xeon(R) Platinum 8468V CPU family: 6 Model: 143 Thread(s) per core: 2 Core(s) per socket: 48 Socket(s): 2 Stepping: 8 CPU(s) scaling MHz: 70% CPU max MHz: 3800.0000 CPU min MHz: 800.0000 BogoMIPS: 4800.00 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf tsc_known_freq pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb cat_l3 cat_l2 cdp_l3 invpcid_single intel_ppin cdp_l2 ssbd mba ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm cqm rdt_a avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb intel_pt avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local split_lock_detect avx_vnni avx512_bf16 wbnoinvd dtherm ida arat pln pts hwp hwp_act_window hwp_epp hwp_pkg_req hfi avx512vbmi umip pku ospke waitpkg avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg tme avx512_vpopcntdq rdpid bus_lock_detect cldemote movdiri movdir64b enqcmd fsrm uintr md_clear serialize tsxldtrk pconfig arch_lbr amx_bf16 avx512_fp16 amx_tile amx_int8 flush_l1d arch_capabilities Virtualization: VT-x L1d cache: 4.5 MiB (96 instances) L1i cache: 3 MiB (96 instances) L2 cache: 192 MiB (96 instances) L3 cache: 195 MiB (2 instances) NUMA node(s): 2 NUMA node0 CPU(s): 0-47,96-143 NUMA node1 CPU(s): 48-95,144-191 Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Not affected Vulnerability Retbleed: Not affected Vulnerability Spec rstack overflow: Not affected Vulnerability Spec store bypass: Mitigation;
https://github.com/vllm-project/vllm/issues/29245
open
[ "usage" ]
2025-11-22T20:41:27Z
2025-12-11T11:23:54Z
3
hucorz
huggingface/candle
3,208
`cudarc` dynamic loading support
Currently, `candle` uses `cudarc` with the `dynamic-linking` feature, which requires the executable to find the DLLs or SOs at startup. However, it would be more convenient if `candle` also supported the `dynamic-loading` feature from `cudarc` to load DLLs or SOs at runtime. Is it possible for `candle` to support it?
https://github.com/huggingface/candle/issues/3208
open
[]
2025-11-22T18:18:25Z
2025-11-25T09:00:27Z
7
mayocream
huggingface/transformers
42,331
SAM3 does not support custom inference resolutions
### System Info Note: I am running the latest git version, sys Info should not be relevant to the issue $ transformers env Traceback (most recent call last): File "/home/master-andreas/panopticon/test_env/bin/transformers", line 3, in <module> from transformers.cli.transformers import main File "/home/master-andreas/panopticon/test_env/lib/python3.12/site-packages/transformers/cli/transformers.py", line 23, in <module> from transformers.cli.serve import Serve File "/home/master-andreas/panopticon/test_env/lib/python3.12/site-packages/transformers/cli/serve.py", line 351, in <module> class Serve: File "/home/master-andreas/panopticon/test_env/lib/python3.12/site-packages/transformers/cli/serve.py", line 658, in Serve ) -> ChatCompletionChunk: ^^^^^^^^^^^^^^^^^^^ NameError: name 'ChatCompletionChunk' is not defined ### Who can help? @yonigozlan ### Information - [ ] The official example scripts - [x] My own modified scripts ### Tasks - [x] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...) - [ ] My own task or dataset (give details below) ### Reproduction ```py """ Test script for SAM3 text prompting only. This script demonstrates how to use SAM3 for text-based segmentation on images. """ import torch from PIL import Image import requests from transformers import Sam3Processor, Sam3Model import os INFERENCE_RESOLUTION = (1008, 1008) # If run with anything else other than 1008 it fails # INFERENCE_RESOLUTION = (1400, 1400) def test_sam3_text_prompting(): """Test SAM3 with text prompting on a sample image.""" # Set device device = "cpu" print(f"Using device: {device}") # Load model and processor print("Loading SAM3 model and processor...") model = Sam3Model.from_pretrained("facebook/sam3").to(device) processor = Sam3Processor.from_pretrained("facebook/sam3") # Load a sample image print("Loading sample image...") image_url = "http://images.cocodataset.org/val2017/000000077595.jpg" image = Image.open(requests.get(image_url, stream=True).raw).convert("RGB") # Define text prompts to test text_prompts = ["cat", "ear", "eye"] for text_prompt in text_prompts: print(f"\nTesting text prompt: '{text_prompt}'") # Prepare inputs inputs = processor(images=image, text=text_prompt, size=INFERENCE_RESOLUTION, return_tensors="pt").to(device) # Run inference with torch.no_grad(): outputs = model(**inputs) # Post-process results results = processor.post_process_instance_segmentation( outputs, threshold=0.5, mask_threshold=0.5, target_sizes=inputs.get("original_sizes").tolist() )[0] # Display results num_objects = len(results['masks']) print(f"Found {num_objects} objects matching '{text_prompt}'") if num_objects > 0: # Show scores for first few objects scores = results['scores'] print(f"Confidence scores: {scores[:min(3, len(scores))].tolist()}") # Show bounding boxes for first object if 'boxes' in results and len(results['boxes']) > 0: box = results['boxes'][0] print(f"First object bounding box (xyxy): {box.tolist()}") if __name__ == "__main__": print("SAM3 Text Prompting Test Script") print("=" * 40) try: test_sam3_text_prompting() print("\n✓ All tests completed successfully!") except Exception as e: print(f"\n✗ Test failed with error: {e}") raise ``` Output when INFERENCE_RESOLUTION=[1400, 1400]: ```sh $ py test_sam3_text.py SAM3 Text Prompting Test Script ======================================== Using device: cpu Loading SAM3 model and processor... Loading weights: 100%|█| 1468/1468 [00:00<00:00, 2709.52it/s, Materializing param=vision_encoder.neck.fpn Loading sample image... Testing text prompt: 'cat' ✗ Test failed with error: The size of tensor a (10000) must match the size of tensor b (5184) at non-singleton dimension 2 Traceback (most recent call last): File "/home/master-andreas/panopticon/test_sam3_text.py", line 124, in <module> test_sam3_text_prompting() File "/home/master-andreas/panopticon/test_sam3_text.py", line 48, in test_sam3_text_prompting outputs = model(**inputs) ^^^^^^^^^^^^^^^ File "/home/master-andreas/panopticon/test_env/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl return self._call_impl(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/master-andreas/panopticon/test_env/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl return forward_call(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/master-andreas/panopticon/test_env/lib/python3.12/site-packages/transformers/utils/generic.py", line 938, in wrapper
https://github.com/huggingface/transformers/issues/42331
closed
[ "bug" ]
2025-11-21T22:17:08Z
2025-12-10T22:46:39Z
3
Kallinteris-Andreas
huggingface/lerobot
2,500
question about the gr00t policy
hi, I see here https://huggingface.co/docs/lerobot/en/groot that gr00t is intergrated into lerobot. is it in sync with the original repo: https://github.com/NVIDIA/Isaac-GR00T ? I see in original repo that the dataset used to fine-tune, is a bit different from the original lerobot format, like libero dataset (https://huggingface.co/datasets/physical-intelligence/libero) used in pi model , therefore i wonder what dataset format should be used here in lerbot policy training ? any example dataset that is passed to `--dataset.repo_id=$DATASET_ID` ? is it a post-processed dataset ?
https://github.com/huggingface/lerobot/issues/2500
open
[ "question", "policies" ]
2025-11-21T21:45:19Z
2025-12-03T14:03:34Z
null
yanan1116
vllm-project/vllm
29,192
Tool Calling Parsers Fail to Populate tool_calls Array for Qwen2.5-Coder Models
# Tool Calling Parsers Fail to Populate `tool_calls` Array for Qwen2.5-Coder Models ## Environment - **vLLM Version**: v0.11.2.dev115+g56669c1f2 (Blackwell build) - **Model**: Qwen/Qwen2.5-Coder-14B-Instruct-AWQ - **Quantization**: AWQ - **Python Version**: 3.x (Docker container) - **GPU**: NVIDIA GeForce RTX 5080 (16GB, Blackwell/sm_120) - **Platform**: WSL2, Linux 6.6.87.2-microsoft-standard-WSL2 ## Description When using tool calling with Qwen2.5-Coder models, the model correctly generates tool calls in `<tools>` XML format, but both `qwen3_xml` and `qwen3_coder` parsers fail to extract these tool calls into the `tool_calls` array in the API response. The tool call information remains in the `content` field but the `tool_calls` array stays empty. ## Steps to Reproduce 1. Start vLLM with Qwen2.5-Coder and tool calling parser: ```bash python -m vllm.entrypoints.openai.api_server \ --model Qwen/Qwen2.5-Coder-14B-Instruct-AWQ \ --quantization awq \ --enable-auto-tool-choice \ --tool-call-parser qwen3_xml # or qwen3_coder ``` 2. Send a tool calling request: ```bash curl -s http://localhost:8002/v1/chat/completions \ -H "Content-Type: application/json" \ -d '{ "model": "qwen2.5-coder-14b-awq", "messages": [{"role": "user", "content": "What is the weather in San Francisco?"}], "tools": [ { "type": "function", "function": { "name": "get_weather", "description": "Get the current weather for a location", "parameters": { "type": "object", "properties": { "location": { "type": "string", "description": "The city and state, e.g. San Francisco, CA" } }, "required": ["location"] } } } ], "tool_choice": "auto" }' ``` ## Actual Output ```json { "id": "chatcmpl-xxx", "object": "chat.completion", "model": "qwen2.5-coder-14b-awq", "choices": [ { "message": { "role": "assistant", "content": "<tools>\n{\n \"name\": \"get_weather\",\n \"arguments\": {\n \"location\": \"San Francisco, CA\"\n }\n}\n</tools>", "tool_calls": [] } } ] } ``` ## Expected Output ```json { "id": "chatcmpl-xxx", "object": "chat.completion", "model": "qwen2.5-coder-14b-awq", "choices": [ { "message": { "role": "assistant", "content": "", "tool_calls": [ { "type": "function", "id": "call_0", "function": { "name": "get_weather", "arguments": "{\"location\": \"San Francisco, CA\"}" } } ] } } ] } ``` ## Analysis ### Model Output (Correct) The model correctly generates tool calls in the expected `<tools>` XML format: ```xml <tools> { "name": "get_weather", "arguments": { "location": "San Francisco, CA" } } </tools> ``` ### Parser Behavior (Incorrect) Both recommended parsers fail to extract tool calls: - **hermes parser**: Expects `<tool_call>` tags, doesn't match `<tools>` tags - **qwen3_xml parser**: Designed for `<tools>` tags but doesn't populate `tool_calls` array - **qwen3_coder parser**: Also designed for Qwen but fails to populate array ### Root Cause The parsers appear to load correctly (visible in logs as `'tool_call_parser': 'qwen3_xml'`) but the extraction logic fails to populate the OpenAI-compatible `tool_calls` array structure. ## Workaround Manual extraction from the `content` field: ```python import re import json def extract_tool_calls(response): """Extract tool calls from Qwen2.5-Coder <tools> tags""" content = response['choices'][0]['message']['content'] pattern = r'<tools>\s*({.*?})\s*</tools>' match = re.search(pattern, content, re.DOTALL) if match: tool_data = json.loads(match.group(1)) return [{ "type": "function", "function": { "name": tool_data["name"], "arguments": json.dumps(tool_data["arguments"]) } }] return [] ``` ## Additional Context ### Multi-AI Consultation Results Consulted with multiple AI models for parser recommendation: - **Qwen3 Coder (480B)**: Recommended `qwen3_xml` parser - **DeepSeek V3.1**: Ranked `qwen3_xml` (90% confidence), `qwen3_coder` (80% confidence) - **Claude Sonnet 4.5**: Confirmed tag mismatch between Hermes and Qwen formats All models agreed that the parser selection is correct, suggesting the issue is in the parser implementation rather than configuration. ### vLLM Configuration ```python { 'tool_call_parser': 'qwen3_xml', # Confirmed in logs 'enable_auto_tool_choice': True, 'model': 'Qwen/Qwen2.5-Coder-14B-Instruct-AWQ', 'quantization': 'awq', 'max_model_len': 8192 } ``` ## Impact - **Severity**: High - Breaks OpenAI API compatibility for tool calling - **Affected Models**: Likely all Qwen2.5-Coder variants -
https://github.com/vllm-project/vllm/issues/29192
open
[]
2025-11-21T18:31:19Z
2025-11-21T18:31:19Z
0
Platano78
vllm-project/vllm
29,180
[Bug]: Recorded `EngineCoreEventType.QUEUED` time is off
### Your current environment <details> </details> ### 🐛 Describe the bug When running benchmarking with the CLI: - on one side the serving point `vllm serve ...` - on the other side the benchmarking client : `vllm bench serve...` (note that the two are running on the same machine, there is no networking delay) I noticed that the `EngineCoreEventType.QUEUED` event recorder on the server side didn't match the time of posting the request. In my understanding these two should events should be approximately equivalent. These values aren't off by a few milliseconds, but here the mismatch can be pretty big, up to a few seconds. I think the reason might be because adding [request to the scheduler](https://github.com/vllm-project/vllm/blob/fcb1d570bb8f95f5b7ded716a52fec902c535f0e/vllm/v1/core/sched/scheduler.py#L1166) cannot be done when the engine is running a decoding or a prefill, see the [`_process_input_queue` function](https://github.com/vllm-project/vllm/blob/fcb1d570bb8f95f5b7ded716a52fec902c535f0e/vllm/v1/engine/core.py#L801), where `add_request()` ultimately gets called. This can introduce delays before the queued event gets recorded, having "floating" requests that are not tracked in the logs. ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29180
closed
[ "bug" ]
2025-11-21T12:58:36Z
2025-11-30T20:56:44Z
4
sducouedic
vllm-project/vllm
29,177
[Usage]: Vllm + Intervl model local infra Image preprocessing / request adding becomes bottleneck even with more CPU cores — how to accelerate?
### Your current environment vllm 0.11.0 ### How would you like to use vllm ### current phenomenon When doing **batched image classification** (64 images per batch) with InternVL3_5-1B, the bottleneck is clearly in the **"Adding requests"** phase (image preprocessing). Even after increasing CPU cores and setting `OMP_NUM_THREADS=16`, the preprocessing speed stays around **50 it/s**, while the actual generation phase is extremely fast (>1500 prompts/s). ```text Adding requests: 100%|██████████| 64/64 [00:01<00:00, 52.67it/s] ← bottleneck Processed prompts: 100%|█| 64/64 [00:00<00:00, 1515.23it/s, est. speed input: 812805.23 tok/s] ``` This means ~95% of the total latency is spent on CPU-side image preprocessing, (I have disabled dynamic resolution) ### Minimal Reproducible Example ```python import os from PIL import Image from vllm import LLM, SamplingParams from transformers import AutoTokenizer model_path = "/data/code/haobang.geng/models/InternVL3_5-1B" tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True) llm = LLM( model=model_path, dtype="bfloat16", max_model_len=4096, gpu_memory_utilization=0.95, limit_mm_per_prompt={"image": 1}, trust_remote_code=True, enforce_eager=False, ) prompt = "<image>\nYou are an image classifier. Output only one word: safe or nsfw." sampling_params = SamplingParams(temperature=0.0, max_tokens=8) batch_inputs = [] for i in range(64): img = Image.open(f"/path/to/images/{i}.jpg").convert("RGB") batch_inputs.append({ "prompt": prompt, "multi_modal_data": {"image": img}, }) outputs = llm.generate(batch_inputs, sampling_params=sampling_params, use_tqdm=True) ``` ### Expected behavior For pure-text batches, Adding requests is >2000 it/s such as qwen3vl. Attempted solutions (all ineffective) ### my attempt to speed up Increase CPU cores / set OMP_NUM_THREADS=16 → no speedup mm_processor_kwargs={"max_dynamic_patch": 1, ...} → seems no speedup Pre-resize images to 384×384 → helps a little (~55 it/s) but still far from ideal ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions. Thank you for the great work on vLLM! Looking forward to a simple way to slove it
https://github.com/vllm-project/vllm/issues/29177
open
[ "usage" ]
2025-11-21T10:56:29Z
2025-12-01T14:08:22Z
3
Passenger12138
huggingface/trl
4,554
Better packing of data with best-fit decrease strategy
Hello, When using packing with the bfd strategy, it looks like too much truncation is done when the seq_length is smaller than the average length of the sequences we want to pack. For example : ```python from datasets import Dataset from trl import pack_dataset examples = { "input_ids": [[1, 2, 3, 4], [5, 6], [7, 8, 9], [10]], "attention_mask": [[1, 1, 1, 1], [1, 0], [1, 0, 0], [1]], } dataset = Dataset.from_dict(examples) packed_dataset = pack_dataset(dataset, seq_length=3, strategy="bfd") print(packed_dataset ) ``` results in: ```python {'input_ids': [[1, 2, 3], [7, 8, 9], [5, 6, 10]], 'attention_mask': [[1, 1, 1], [1, 0, 0], [1, 0, 1]], 'seq_lengths': [[3], [3], [2, 1]]} ``` So the token '4' is missing from the training tokens. In a extreme case: ```python examples_2 = { "input_ids": [[0, 0], [1, 2, 3, 4], [5, 6, 7, 8, 9], [10]], "attention_mask": [[1, 1], [1, 1, 1, 1], [1, 1, 1, 1, 1], [1]], } dataset_2 = Dataset.from_dict(examples_2) print(pack_dataset(dataset_2, seq_length=1, strategy="bfd")[:]) ``` results in: ```python {'input_ids': [[0], [1], [5], [10]], 'attention_mask': [[1], [1], [1], [1]], 'seq_lengths': [[1], [1], [1], [1]]} ``` So here we are basically applying truncation to every sequence instead of having twelve sequences of one token. If we put ourself in a more usefull setting, when I was finetunning on some very long sequences with a seq_lenfth of 4096, the majority of the tokens was discarded y the bfd packing. On my dataset, the bfd method kept only 0.2% of the total training tokens. Is the behavior normal ? I would find it useful to add an option to still have tokens that are deleted in other sequences, even if this is less than ideal. It would be a good compromise between the current versions of bfd and wrapped.
https://github.com/huggingface/trl/issues/4554
closed
[ "✨ enhancement", "❓ question" ]
2025-11-21T07:53:55Z
2025-12-16T20:37:02Z
3
ntnq4
vllm-project/vllm
29,148
[Usage]: Deployment of the embedding models
### Your current environment ```text ============================== System Info ============================== OS : Ubuntu 22.04.5 LTS (x86_64) GCC version : (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 Clang version : Could not collect CMake version : version 3.22.1 Libc version : glibc-2.35 ============================== PyTorch Info ============================== PyTorch version : 2.8.0+cu128 Is debug build : False CUDA used to build PyTorch : 12.8 ROCM used to build PyTorch : N/A ============================== Python Environment ============================== Python version : 3.10.16 (main, Dec 11 2024, 16:24:50) [GCC 11.2.0] (64-bit runtime) Python platform : Linux-5.15.0-161-generic-x86_64-with-glibc2.35 ============================== CUDA / GPU Info ============================== Is CUDA available : True CUDA runtime version : 12.8.61 CUDA_MODULE_LOADING set to : LAZY GPU models and configuration : GPU 0: NVIDIA GeForce RTX 5090 GPU 1: NVIDIA GeForce RTX 5090 GPU 2: NVIDIA GeForce RTX 5090 GPU 3: NVIDIA GeForce RTX 5090 GPU 4: NVIDIA GeForce RTX 5090 GPU 5: NVIDIA GeForce RTX 5090 GPU 6: NVIDIA GeForce RTX 5090 GPU 7: NVIDIA GeForce RTX 5090 Nvidia driver version : 570.172.08 cuDNN version : Probably one of the following: /usr/lib/x86_64-linux-gnu/libcudnn.so.9.7.0 /usr/lib/x86_64-linux-gnu/libcudnn_adv.so.9.7.0 /usr/lib/x86_64-linux-gnu/libcudnn_cnn.so.9.7.0 /usr/lib/x86_64-linux-gnu/libcudnn_engines_precompiled.so.9.7.0 /usr/lib/x86_64-linux-gnu/libcudnn_engines_runtime_compiled.so.9.7.0 /usr/lib/x86_64-linux-gnu/libcudnn_graph.so.9.7.0 /usr/lib/x86_64-linux-gnu/libcudnn_heuristic.so.9.7.0 /usr/lib/x86_64-linux-gnu/libcudnn_ops.so.9.7.0 HIP runtime version : N/A MIOpen runtime version : N/A Is XNNPACK available : True ``` ### How would you like to use vllm When deploying the embedding model, I found that the actual GPU memory usage included not only the model itself but also kv_cache. Is this a reasonable phenomenon? In version v0.9.0, the GPU memory usage was only for the model itself. ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29148
closed
[ "usage" ]
2025-11-21T03:57:59Z
2025-11-21T06:17:18Z
3
Root970103
vllm-project/vllm
29,139
[Feature]: Optimize collectives in TP MoE case using torch.compile pass
### 🚀 The feature, motivation and pitch To avoid redundant work in MoE models in the TP case, sequence parallelism was added to the Deepseek model definition in #24134 and expanded to other models in #24982. However, to avoid performing surgery on the linear layer, the current approach performs more communication than necessary. With a torch.compile custom pass, we can rewrite the graph to remove the redundant computation. ### More details Before the SP optimization, the ops in the model were: ``` - o_proj:[num_tokens, ...] -> [num_tokens, ...] (incomplete results) - all_reduce:[num_tokens, ...] -> [num_tokens, ...] - router:[num_tokens, ...] -> [num_tokens, ...] - experts:[num_tokens, ...] -> [num_tokens, ...] - ... ``` With sequence parallel enabled, this becomes: ``` - o_proj: [num_tokens, ...] -> [num_tokens, ...] (incomplete results) - all_reduce: [num_tokens, ...] -> [num_tokens, ...] - chunk: [num_tokens, ...] -> [num_tokens/tp, ...] - router: [num_tokens/tp, ...] -> [num_tokens/tp, ...] - experts: [num_tokens/tp, ...] -> [num_tokens/tp, ...] - all_gather: [num_tokens/tp, ...] -> [num_tokens, ...] ``` Additionally, experts now properly do the dp+tp<->ep dispatch instead of just the original replicated dp<->ep dispatch. Notice that the `all_reduce` does redundant communication as each TP rank only requires partial results. With a compile pass, we can convert the `all_reduce` -> `chunk` sequence into a `reduce_scatter`: ``` - o_proj: [num_tokens, ...] -> [num_tokens, ...] (incomplete results) - reduce_scatter: [num_tokens, ...] -> [num_tokens/tp, ...] - router: [num_tokens/tp, ...] -> [num_tokens/tp, ...] - experts: [num_tokens/tp, ...] -> [num_tokens/tp, ...] - all_gather: [num_tokens/tp, ...] -> [num_tokens, ...] ``` We should create a new `SequenceParallelismMoEPass`, controlled by a new `PassConfig.enable_sp_moe` flag (following the new naming convention in #27995) so that it can be turned on independently of regular SP. We will likely need to pad the number of tokens to a multiple of TP size, although like described in #29136, there are alternatives. ### Alternatives Alternatively, the original optimization could be done as a compile pass as well, which would significantly clean up the MoE model definitions. However, that would mean that `VLLM_COMPILE` compilation mode would be required for this optimization and if compilation is disabled, the optimization would be disabled as well. Generally we accept lower performance in eager mode as compilation is on by default, but I know there was a reason this was done this way (don't remember why). ### Additional context Original proposal comment: https://github.com/vllm-project/vllm/pull/24982#pullrequestreview-3259494618 cc @tlrmchlsmth @bnellnm @robertgshaw2-redhat @alexm-redhat @zou3519 @nvpohanh @youkaichao ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29139
open
[ "help wanted", "good first issue", "performance", "feature request", "torch.compile" ]
2025-11-21T01:36:06Z
2025-12-07T15:39:48Z
19
ProExpertProg
vllm-project/vllm
29,097
[Docs] Feedback for `/en/latest/`
### 📚 The doc issue no ### Suggest a potential alternative/fix _No response_ ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29097
closed
[ "documentation" ]
2025-11-20T14:53:44Z
2025-11-21T07:51:57Z
2
ch950684-svg
vllm-project/vllm
29,089
[Performance]: Can we use CUDA graph to accelerate the Qwen2_5omniAudioEncoder in Qwen2.5-Omni-3B?
### Proposal to improve performance <img width="3088" height="1264" alt="Image" src="https://github.com/user-attachments/assets/535d7854-b9db-4e40-8f85-1abe08b4d35e" /> The trace graph shows that Qwen2_5omniAudioEncoder has a large number of small kernel startups, indicating significant room for optimization. Can we use CUDA graph to accelerate the Qwen2_5omniAudioEncoder in Qwen2.5-Omni-3B? ### Report of performance regression _No response_ ### Misc discussion on performance _No response_ ### Your current environment (if you think it is necessary) ```text The output of `python collect_env.py` ``` ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29089
open
[ "performance" ]
2025-11-20T12:13:58Z
2025-11-20T12:13:58Z
0
xq25478
vllm-project/vllm
29,078
[Performance]: 多实例导致的cpu占用过高
### Your current environment GPU: RTX4090 cuda version: cuda12.8 vllm version: 0.11.0 中文:我使用triton server的 vllm backend 启动了4个 minerU2.5 模型的实例,我的服务器上有2张卡,我每张卡启动了1个实例,我发现cpu负载有时候极高,几乎占满了我的服务器,我的服务器有96核,vllm backend使用的是AsyncLLMEngine,我观察到在单卡上启动一个实例时,我发送200张小尺寸的文字图做OCR时,fps可以达到最高,也就是每秒可以处理200张的图片,cpu负载在40-50%左右,为了进一步增加性能,我在两张卡上各启动了一个实例,但是我观察到此时cpu负载几乎达到99%,占用了极高的cpu,每个实例的fps只有120左右,性能几乎没有提升。 我做了大量的测试,我开始以为是triton server的问题,但经过排查,我认为问题可能出现在vllm推理时占用了很高的cpu,因为我不使用triton server,使用 `vllm serve`来模拟同样的情况,每个vllm实例推理时也占用掉了20-30%的cpu,如果这样,我的服务器即使有再多的GPU,也不能够提升模型的性能,我该如何调试? english: I launched 4 instances of the minerU2.5 model using the vllm backend of Triton Server. My server is equipped with 2 GPUs, with 1 instance running on each GPU. However, I noticed that the CPU load sometimes spikes to extremely high levels, nearly maxing out the server—which has 192 CPU cores. The vllm backend uses AsyncLLMEngine. When running a single instance on one GPU and sending 200 small-sized text images for OCR, I achieved the highest FPS—processing up to 200 images per second—with the CPU load hovering around 40-50%. To further improve performance, I launched one instance on each of the two GPUs. But in this scenario, the CPU load reached nearly 99% (extremely high usage), and each instance only achieved around 120 FPS, with almost no performance gain. I conducted numerous tests. Initially, I suspected the issue was with Triton Server, but after troubleshooting, I believe the problem lies in the high CPU usage during vllm inference. Even when not using Triton Server—simulating the same scenario with `vllm serve`—each vllm instance consumes 20-30% of the CPU. If this persists, adding more GPUs to the server will not improve model performance. How should I debug this? ### How would you like to use vllm I want to run inference of a [[MinerU2.5-2509-1.2B]().](https://huggingface.co/opendatalab/MinerU2.5-2509-1.2B) I don't know how to integrate it with vllm. ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29078
closed
[ "usage" ]
2025-11-20T08:26:35Z
2025-11-21T02:17:51Z
4
zjq1996518
huggingface/transformers
42,291
Can we disable IPython progress bar and use normal tqdm bar?
I like the normal tqdm bar much better, it is lighter, cleaner, simpler, and less stress on my eyes (no green color). I would love to have an option to use tqdm bar and not IPython bar.
https://github.com/huggingface/transformers/issues/42291
closed
[]
2025-11-20T01:26:11Z
2025-12-28T08:02:45Z
1
weathon
vllm-project/vllm
29,023
[Feature]: Disable logging `/metrics`
### 🚀 The feature, motivation and pitch - IGW hits `/metrics` continuously to understand the current load on the system - This leads to an overload of logs - We can disable this with `--disable-uvicorn-access-log`, but lose access to all access logs We should have `--disable-uvicorn-metrics-access-log` to avoid logging * just * metrics. Per Gemini, we can do this with something like: ```python # Define the routes for which access logs should be disabled EXCLUDE_PATHS = ["/health", "/metrics"] class EndpointFilter(logging.Filter): def filter(self, record: logging.LogRecord) -> bool: # Check if the log record contains arguments and if the path matches an excluded path if record.args and len(record.args) >= 3: path = record.args[2] # The path is typically the third argument in uvicorn access logs if path in EXCLUDE_PATHS: return False # Exclude this log record return True # Include all other log records ``` Create a command line arg like `--disable-uvicorn-metrics-access-log`which selectively disables logging hits to `/metrics` ### Alternatives _No response_ ### Additional context _No response_ ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/29023
open
[ "help wanted", "good first issue", "feature request" ]
2025-11-19T18:25:48Z
2025-11-19T21:57:34Z
5
robertgshaw2-redhat
huggingface/sentence-transformers
3,575
How to override model's `max_seq_length`?
It seems that impossible to override model's max length from `sentence_bert_config.json`. ```python from sentence_transformers import SentenceTransformer m = SentenceTransformer("intfloat/e5-small", tokenizer_kwargs={"model_max_length":3}) print(m.tokenize(["hi hi hi hi hi hi hi hi hi hi hi hi hi"])) # {'input_ids': tensor([[ 101, 7632, 7632, 7632, 7632, 7632, 7632, 7632, 7632, 7632, 7632, 7632, # 7632, 7632, 102]]), 'token_type_ids': tensor([[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]), 'attention_mask': tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]])} print(m.tokenize(["hi hi hi hi hi hi hi hi hi hi hi hi hi"], truncation=True)) # {'input_ids': tensor([[ 101, 7632, 7632, 7632, 7632, 7632, 7632, 7632, 7632, 7632, 7632, 7632, # 7632, 7632, 102]]), 'token_type_ids': tensor([[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]), 'attention_mask': tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]])} print(m[0].tokenizer(["hi hi hi hi hi hi hi hi hi hi hi hi hi"], truncation=True)) # {'input_ids': [[101, 7632, 102]], 'token_type_ids': [[0, 0, 0]], 'attention_mask': [[1, 1, 1]]} m.max_seq_length = 3 print(m.tokenize(["hi hi hi hi hi hi hi hi hi hi hi hi hi"])) # {'input_ids': tensor([[ 101, 7632, 102]]), 'token_type_ids': tensor([[0, 0, 0]]), 'attention_mask': tensor([[1, 1, 1]])} ``` This is happening because during load it load `max_seq_length` from `sentence_bert_config` and then in `Transformers` it will override `max_seq_length` only it wasn't set in `sentence_bert_config` https://github.com/huggingface/sentence-transformers/blob/ad28c0a982acc39c73abdf0019faca10f227ef28/sentence_transformers/models/Transformer.py#L101-L118 even if `model_max_length` is passed in `tokenizer_kwargs` and then `max_seq_length` will be used as `max_length` instead of passed in kwargs https://github.com/huggingface/sentence-transformers/blob/ad28c0a982acc39c73abdf0019faca10f227ef28/sentence_transformers/models/Transformer.py#L319-L327 Probably this can be fixed by ```diff max_seq_length = min(max_seq_length, self.tokenizer.model_max_length) ``` Source https://github.com/embeddings-benchmark/mteb/pull/3587#discussion_r2542434603 I think this is cause of https://github.com/huggingface/sentence-transformers/issues/3187
https://github.com/huggingface/sentence-transformers/issues/3575
open
[]
2025-11-19T16:42:27Z
2025-11-20T13:47:13Z
null
Samoed
huggingface/trl
4,546
Does TRL support PipelineRL for compute efficiency?
Hi 👋, I'm trying to understand whether TRL currently supports (or plans to support) the PipelineRL approach described here: - Paper: [https://arxiv.org/pdf/2509.19128v2](https://arxiv.org/pdf/2509.19128v2?utm_source=chatgpt.com) - Overview: [https://arxiv.org/html/2509.19128](https://arxiv.org/html/2509.19128?utm_source=chatgpt.com) PipelineRL introduces an actor–learner pipeline with in-flight weight updates, where actors keep generating while the learner updates weights concurrently. This reduces policy lag and improves GPU utilization for long-context RL runs. Does TRL currently support this kind of pipelineRL workflow, or is there a recommended way to approximate it using the existing TRL trainers (GRPO + vLLM)? If not, I'd love suggestions or best practices for building something similar on top of TRL. Thanks! 🙏
https://github.com/huggingface/trl/issues/4546
open
[ "✨ enhancement", "❓ question" ]
2025-11-19T12:39:29Z
2025-11-22T12:43:54Z
3
harisarang
vllm-project/vllm
28,996
[Usage]: How to run a single data parallel deployment across multiple nodes without ray
### Your current environment 2 Nodes, each node has 8 H20 GPUs. ### How would you like to use vllm According to https://docs.vllm.ai/en/latest/serving/data_parallel_deployment/#internal-load-balancing ```shell # node0 vllm serve Qwen3-Coder-480B-A35B-Instruct --trust-remote-code --max-num-seqs 64 --max-model-len 131072 --port $PORT0 --host :: --data-parallel-size 2 --data-parallel-size-local 1 --data-parallel-address $NODE0_IPV6 --data-parallel-rpc-port $PORT1 # node1 vllm serve Qwen3-Coder-480B-A35B-Instruct --trust-remote-code --max-num-seqs 64 --max-model-len 131072 --headless --data-parallel-size 2 --data-parallel-size-local 1 --data-parallel-start-rank 1 --data-parallel-address $NODE0_IPV6 --data-parallel-rpc-port $NODE0_PORT1 ``` but all of them are hanging on waiting for init message from front-end. ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/28996
closed
[ "usage" ]
2025-11-19T06:47:22Z
2025-11-27T06:17:22Z
3
crystalww
vllm-project/vllm
28,986
[Feature]: Fused Kernel for GPT-OSS Router
### 🚀 The feature, motivation and pitch <img width="1257" height="250" alt="Image" src="https://github.com/user-attachments/assets/31eba061-522c-4521-b0a9-9f25bb36c3df" /> - Right now, we spend ~3.5% of the layer in the expert selection - The operation is unfused Write a fused kernel like we have for deepseek grouped_topk ### Alternatives - torch compile - triton - cuda ### Additional context _No response_ ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/28986
open
[ "help wanted", "good first issue", "feature request" ]
2025-11-19T03:18:25Z
2025-12-12T16:16:37Z
7
robertgshaw2-redhat
huggingface/transformers.js
1,458
ONNX Backend Env variable
### Question Hi, For some context, I'm building an application that uses some of the models on huggingface as an annotation tool that helps create annotations for training a specialised model. As for the specialised model, I am able to export them to onnx, and I was able to run this model in the same application, but I have to manually install the same onnxruntime-web version to be able to do so. I looked into the docs [here](https://huggingface.co/docs/transformers.js/api/backends/onnx#module_backends/onnx.createInferenceSession), but I cannot access these functions through `env.backends.onnx`. I've tried `console.log(env.backends.onnx.isONNXProxy())` and got ``` Uncaught (in promise) TypeError: env.backends.onnx.isONNXProxy is not a function ``` Is there a way I can access the same inference session through this package? --------------------------------- My `package.json` ``` { "dependencies": { "@huggingface/transformers": "3.7.5", "onnxruntime-web": "1.22.0-dev.20250409-89f8206ba4" }, } ```
https://github.com/huggingface/transformers.js/issues/1458
open
[ "question" ]
2025-11-19T01:26:02Z
2025-11-25T15:36:13Z
null
Heinrik-20
vllm-project/vllm
28,956
[Bug]: OOM when profiling multimodal model with multiple images
### Your current environment vLLM 0.11.0 ### 🐛 Describe the bug As per title. The error log is as follows: ``` [multiproc_executor.py:671] Traceback (most recent call last): [multiproc_executor.py:671] File "/root/miniconda3/lib/python3.11/site-packages/vllm/v1/executor/multiproc_executor.py", line 666, in worker_busy_loop [multiproc_executor.py:671] output = func(*args, **kwargs) [multiproc_executor.py:671] ^^^^^^^^^^^^^^^^^^^^^ [multiproc_executor.py:671] File "/root/miniconda3/lib/python3.11/site-packages/torch/utils/_contextlib.py", line 120, in decorate_context [multiproc_executor.py:671] return func(*args, **kwargs) [multiproc_executor.py:671] ^^^^^^^^^^^^^^^^^^^^^ [multiproc_executor.py:671] File "/root/miniconda3/lib/python3.11/site-packages/vllm/v1/worker/gpu_worker.py", line 263, in determine_available_memory [multiproc_executor.py:671] self.model_runner.profile_run() [multiproc_executor.py:671] File "/root/miniconda3/lib/python3.11/site-packages/vllm/v1/worker/gpu_model_runner.py", line 3379, in profile_run [multiproc_executor.py:671] expanded = output.new_zeros( [multiproc_executor.py:671] ^^^^^^^^^^^^^^^^^ [multiproc_executor.py:671] torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 3.00 GiB. GPU 6 has a total capacity of 139.81 GiB of which 2.58 GiB is free. Including non-PyTorch memory, this process has 137.21 GiB memory in use. Of the allocated memory 134.77 GiB is allocated by PyTorch, and 255.64 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) ``` Looks like we only need **ONE** encoder cache with shape `(encoder_budget, encoder_output_shape[-1])` rather than `len(dummy_encoder_outputs)` ones. https://github.com/vllm-project/vllm/blob/da8dadf68b5a2af849e7c5fd35ce9b8525d8d398/vllm/v1/worker/gpu_model_runner.py#L4128-L4144 ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/28956
closed
[ "bug" ]
2025-11-18T17:36:55Z
2025-11-25T12:38:37Z
7
imShZh
huggingface/lerobot
2,475
Why there is difference between async inference and local inference in image resize?
I read code between `src/lerobot/async_inference/policy_server.py` and `src/lerobot/scripts/lerobot_record.py`. I found difference in these 2 code about inference which causes different image shape 1. `src/lerobot/scripts/lerobot_record.py` use this to deal with observation And `prepare_observation_for_inference` is like this: ```python def prepare_observation_for_inference( observation: dict[str, np.ndarray], device: torch.device, task: str | None = None, robot_type: str | None = None, ) -> RobotObservation: for name in observation: observation[name] = torch.from_numpy(observation[name]) if "image" in name: observation[name] = observation[name].type(torch.float32) / 255 observation[name] = observation[name].permute(2, 0, 1).contiguous() observation[name] = observation[name].unsqueeze(0) observation[name] = observation[name].to(device) observation["task"] = task if task else "" observation["robot_type"] = robot_type if robot_type else "" return observation ``` Here no **resize** operation in images i think. 2. in async_inference policy_server.py,it uses But here function`prepare_raw_observation` makes sense on image shape ```python def prepare_raw_observation( robot_obs: RawObservation, lerobot_features: dict[str, dict], policy_image_features: dict[str, PolicyFeature], ) -> Observation: """Matches keys from the raw robot_obs dict to the keys expected by a given policy (passed as policy_image_features).""" # 1. {motor.pos1:value1, motor.pos2:value2, ..., laptop:np.ndarray} -> # -> {observation.state:[value1,value2,...], observation.images.laptop:np.ndarray} lerobot_obs = make_lerobot_observation(robot_obs, lerobot_features) # 2. Greps all observation.images.<> keys image_keys = list(filter(is_image_key, lerobot_obs)) # state's shape is expected as (B, state_dim) state_dict = {OBS_STATE: extract_state_from_raw_observation(lerobot_obs)} image_dict = { image_k: extract_images_from_raw_observation(lerobot_obs, image_k) for image_k in image_keys } # Turns the image features to (C, H, W) with H, W matching the policy image features. # This reduces the resolution of the images image_dict = { key: resize_robot_observation_image(torch.tensor(lerobot_obs[key]), policy_image_features[key].shape) for key in image_keys } if "task" in robot_obs: state_dict["task"] = robot_obs["task"] return {**state_dict, **image_dict} ``` Here the shape of observation images is modified to policy config ```python def resize_robot_observation_image(image: torch.tensor, resize_dims: tuple[int, int, int]) -> torch.tensor: assert image.ndim == 3, f"Image must be (C, H, W)! Received {image.shape}" # (H, W, C) -> (C, H, W) for resizing from robot obsevation resolution to policy image resolution image = image.permute(2, 0, 1) dims = (resize_dims[1], resize_dims[2]) # Add batch dimension for interpolate: (C, H, W) -> (1, C, H, W) image_batched = image.unsqueeze(0) # Interpolate and remove batch dimension: (1, C, H, W) -> (C, H, W) resized = torch.nn.functional.interpolate(image_batched, size=dims, mode="bilinear", align_corners=False) return resized.squeeze(0) ``` I found this when I can inference correctly locally by made weird action outputs from async inference. it must be caused by my didn't resize input image when training. -.- version:deb9596bd3796c03ae3a5a6b81b63c1dba296256
https://github.com/huggingface/lerobot/issues/2475
open
[ "question" ]
2025-11-18T14:32:17Z
2025-11-24T02:23:13Z
null
milong26
vllm-project/vllm
28,943
[Usage]: what's the right way to run embedding model in vllm 0.11.0
### Your current environment ```text The output of `python collect_env.py` ``` in vllm 0.8.7,I use following code to run local vllm,all is right: ``` self.engine_args = EngineArgs( model=self.model_path, dtype='half', task="embed", trust_remote_code=True, limit_mm_per_prompt={"image": 1}, ) e = asdict(self.engine_args) self.max_len = 100 self.llm = LLM(**e) out = self.llm.embed(datas) ``` But in vllm 0.11.0 according to the document https://www.aidoczh.com/vllm/models/pooling_models.html,it use runner=='pooling' to run embedding task. What's the diffenence? Could the 'task' arg 'embed' still take effect? ### How would you like to use vllm I want to run inference of a [specific model](put link here). I don't know how to integrate it with vllm. ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/28943
open
[ "usage" ]
2025-11-18T13:47:57Z
2025-11-20T10:49:12Z
3
neverneverendup
huggingface/trl
4,541
Is attn_implementation=sdpa not supported when using SFTTrainer with mllama?
When trying to use `sdpa` with mllama I get an error using the default collator. Upon writing my own collator it works. When using `eager` implementation it gives cuda oom error. Is `sdpa` not supported?
https://github.com/huggingface/trl/issues/4541
open
[]
2025-11-18T11:57:01Z
2025-11-18T11:57:01Z
0
osaidr
vllm-project/vllm
28,930
[Usage]: How to build a qwen3vl embedding model with a custom mlp layer on the top use vllm?
### Your current environment ```text The output of `python collect_env.py` ``` Hi friends! I train a sft model built upon qwen3vl 2b model, we put a mlp layer on it to compress the embedding size of the backbone model. Now I want to use vllm 0.11.0 to serve it but I meet some confuse. Here is my custom class code ``` from argparse import Namespace from dataclasses import asdict from typing import Literal, NamedTuple, Optional, TypedDict, Union, get_args import torch import torch.nn as nn from vllm.model_executor.models.qwen3_vl import Qwen3VLForConditionalGeneration from vllm.v1.pool.metadata import PoolingMetadata from vllm.v1.sample.metadata import SamplingMetadata from vllm.config import VllmConfig from vllm.multimodal import MULTIMODAL_REGISTRY class CustomQwenVL3BPool(nn.Module): def __init__( self ): super().__init__() self.out = torch.nn.Sequential( torch.nn.Linear(2048, 512), torch.nn.SiLU(), torch.nn.Linear(512, 128) ) def get_prompt_lens(self, hidden_states: Union[torch.Tensor, list[torch.Tensor]], pooling_metadata: PoolingMetadata, ) -> torch.Tensor: return pooling_metadata.prompt_lens def forward( self, hidden_states: torch.Tensor, pooling_metadata: PoolingMetadata, ) -> Union[list[torch.Tensor], torch.Tensor]: # 1 提取lasttoken prompt_lens = self.get_prompt_lens(hidden_states, pooling_metadata) last_token_flat_indices = torch.cumsum(prompt_lens, dim=0) - 1 hidden_states = hidden_states[last_token_flat_indices] # 2 mlp压缩维度 mlp_output = self.out(hidden_states) # 3 正则化输出,需要check下vllm是否会再次norm normalized_output = F.normalize(mlp_output, p=2, dim=-1) return normalized_output class CustomQwen3VLForConditionalGeneration(Qwen3VLForConditionalGeneration): def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""): super().__init__(vllm_config=vllm_config, prefix=prefix) self._pooler = CustomQwenVL3BPool() ``` When I run above code using local mode of vllm , error log says **"[adapters.py:79] ST projector loading failed".Does anybody know why?** BTW,what's the best practice to make a custom embedding model with mlp in vllm 0.11.0 ### How would you like to use vllm I want to run inference of a [specific model](put link here). I don't know how to integrate it with vllm. ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/28930
closed
[ "usage" ]
2025-11-18T10:32:07Z
2025-12-23T04:49:30Z
10
neverneverendup
vllm-project/vllm
28,929
[Usage]: How
=
https://github.com/vllm-project/vllm/issues/28929
closed
[ "usage" ]
2025-11-18T10:26:17Z
2025-11-18T10:30:53Z
0
neverneverendup
huggingface/datasets
7,869
Why does dataset merge fail when tools have different parameters?
Hi, I have a question about SFT (Supervised Fine-tuning) for an agent model. Suppose I want to fine-tune an agent model that may receive two different tools: tool1 and tool2. These tools have different parameters and types in their schema definitions. When I try to merge datasets containing different tool definitions, I get the following error: TypeError: Couldn't cast array of type struct<refundFee: struct<description: string, type: string>, ... , servicerId: struct<description: string, type: string>> to { 'refundFee': {'description': Value(dtype='string'), 'type': Value(dtype='string')}, ... 'templateId': {'description': Value(dtype='string'), 'type': Value(dtype='string')} } From my understanding, the merge fails because the tools column's nested structure is different across datasets — e.g., one struct contains an extra field servicerId while the other does not. This causes HuggingFace Datasets (and its underlying Apache Arrow schema) to reject the merge. My question is: why is it designed this way? Is this strict schema matching a hard requirement of the library? Is there a recommended way to merge datasets with different tool schemas (different parameters and types)? For an agent model supporting multiple tools, what's the best practice for preparing/merging training data without losing flexibility? Any guidance or design rationale would be greatly appreciated. Thanks!
https://github.com/huggingface/datasets/issues/7869
open
[]
2025-11-18T08:33:04Z
2025-11-30T03:52:07Z
1
hitszxs
vllm-project/vllm
28,903
[Bug]: vllm inference on qwen3-vl when use_upstream_fa is False
### Your current environment pip show torch vllm flash-attn Name: torch Version: 2.8.0 --- Name: vllm Version: 0.11.0 Name: flash_attn Version: 2.8.3 ### 🐛 Describe the bug unit-test code as the follows, when simple qwen3-0.6B can run; but qwen3-vl-4b not run ```python #coding=utf-8 """ 写单元测试来验证FA和VLLM的可用性和兼容性 """ import torch from flash_attn import flash_attn_func import unittest import vllm # from vllm.attention.backends import get_attn_backend class TestFA_VLLM(unittest.TestCase): def testFA(self,): # 检查CUDA是否可用及设备 print(f"CUDA available: {torch.cuda.is_available()}") print(f"Current device: {torch.cuda.current_device()}") print(f"Device name: {torch.cuda.get_device_name()}") # 尝试创建一个简单的张量并移动到GPU try: q = torch.randn(1, 1, 16, 64, dtype=torch.float16, device='cuda') k = torch.randn(1, 1, 16, 64, dtype=torch.float16, device='cuda') v = torch.randn(1, 1, 16, 64, dtype=torch.float16, device='cuda') output = flash_attn_func(q, k, v) print("FlashAttention test passed!") except Exception as e: print(f"FlashAttention test failed: {e}") def oriTestVLLM(self,): # 打印当前使用的attention后端 print("Available CUDA devices:", torch.cuda.device_count()) print("Current device:", torch.cuda.current_device()) print("Device name:", torch.cuda.get_device_name()) # 检查vLLM配置 print("vLLM version:", vllm.__version__) # 尝试创建一个小模型来触发后端初始化 try: from vllm import LLM llm = LLM(model="Qwen/Qwen3-0.6B", max_model_len=256) print("vLLM初始化成功!") prompt = "这是一个测试提示。" response = llm.generate(prompt) print("rollout测试成功! 生成的文本:", response) except Exception as e: print(f"vLLM初始化失败: {e}") def testVLLM(self,): # 打印当前使用的attention后端 print("Available CUDA devices:", torch.cuda.device_count()) print("Current device:", torch.cuda.current_device()) print("Device name:", torch.cuda.get_device_name()) # 尝试创建一个小模型来触发后端初始化 try: MODEL_PATH = "Qwen/Qwen3-VL-4B-Instruct" from vllm import LLM from vllm import LLM, SamplingParams from vllm.assets.image import ImageAsset # vLLM 内置工具,帮你把路径 → PIL from vllm.assets.video import VideoAsset # 如果以后想加视频同理 # 随便用一张图就行 image_path = "" from PIL import Image image = Image.open(image_path) # 方式 B:URL # image = ImageAsset("image", "https://xxx.jpg").pil_image # Qwen3-VL 要求的对话模板 messages = [ { "role": "user", "content": [ {"type": "image", "image": image}, # 图像字段 {"type": "text", "text": "请描述这张图片。"} ] } ] # 用 transformers 的 apply_chat_template 把 messages → 模型输入 from transformers import AutoTokenizer tok = AutoTokenizer.from_pretrained(MODEL_PATH) prompt = tok.apply_chat_template( messages, tokenize=False, add_generation_prompt=True ) # ---------- ④ 生成 ---------- sampling_params = SamplingParams( temperature=0.7, max_tokens=512, stop_token_ids=[tok.eos_token_id, tok.convert_tokens_to_ids("<|im_end|>")] ) llm = LLM(model=MODEL_PATH, max_model_len=4096, limit_mm_per_prompt={"image": 1, "video": 0}, # 每张 prompt 最多 1 张图 dtype="bfloat16", # A100/H100 可开;消费卡用 "float16" gpu_memory_utilization=0.9,) print("vLLM初始化成功!") outputs = llm.generate( {"prompt": prompt, "multi_modal_data": {"image": image}}, # 关键:把图也传进去 sampling_params=sampling_params ) response = outputs[0].outputs[0].text print("rollout测试成功! 生成的文本:", response) except Exception as e: print(f"vLLM初始化失败: {e}") if __name__ == "__main__": unittest.main() ``` error is :vllm/vllm_flash_attn/flash_attn_interface.py", line 233, in flash_attn_varlen_func [rank0]: out, softmax_lse = torch.ops._vllm_fa2_C.varlen_fwd( [rank0]: File "/usr/local/lib/python3.10/dist-packages/torch/_ops.py", line 1243, in __call__ [rank0]: return self._op(*args, **kwargs) [rank0]: torch.AcceleratorError: CUDA error: the provided PTX was compiled with an unsupported toolchain. Then, I review the code in https://github.com/vllm-project/vllm/blob/main/vllm/model_executor/models/qwen3_vl.py#L375 it default set use_upstream_fa = False, when I change it to True, it works? the vllm version is 0.11.0 ### Before submitting a new issue... -
https://github.com/vllm-project/vllm/issues/28903
closed
[ "bug" ]
2025-11-18T03:54:11Z
2025-11-18T08:18:09Z
1
hedes1992
huggingface/lerobot
2,465
loss:nan grdn:nan How to solve the gradient explosion problem in PI05 training?
When training Pi05 using Lerobot, has anyone encountered a situation where gradients explode immediately after training? Errors occur when the batch_size is set to 64 or 32. How can this be resolved? Below are my training commands and error logs. python src/lerobot/scripts/lerobot_train.py --dataset.repo_id=aa_merged280 --policy.type=pi05 \ --output_dir=./outputs/pi05_training2 --job_name=pi05_training2 \ --policy.pretrained_path=lerobot/pi05_base --policy.compile_model=true \ --policy.gradient_checkpointing=true --wandb.enable=true --policy.dtype=bfloat16 \ --steps=100000 --policy.device=cuda --batch_size=32 --policy.push_to_hub=false INFO 2025-11-17 22:07:40 ot_train.py:351 step:200 smpl:6K ep:9 epch:0.03 loss:nan grdn:nan lr:2.5e-06 updt_s:4.478 data_s:0.038 WARNING 2025-11-17 22:07:40 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper. INFO 2025-11-17 22:22:38 ot_train.py:351 step:400 smpl:13K ep:18 epch:0.06 loss:nan grdn:nan lr:7.5e-06 updt_s:4.458 data_s:0.022 WARNING 2025-11-17 22:22:38 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper. INFO 2025-11-17 22:37:34 ot_train.py:351 step:600 smpl:19K ep:27 epch:0.10 loss:nan grdn:nan lr:1.3e-05 updt_s:4.456 data_s:0.022 WARNING 2025-11-17 22:37:34 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper. INFO 2025-11-17 22:52:31 ot_train.py:351 step:800 smpl:26K ep:36 epch:0.13 loss:nan grdn:nan lr:1.8e-05 updt_s:4.456 data_s:0.022 WARNING 2025-11-17 22:52:31 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper. INFO 2025-11-17 23:07:29 ot_train.py:351 step:1K smpl:32K ep:45 epch:0.16 loss:nan grdn:nan lr:2.3e-05 updt_s:4.459 data_s:0.022
https://github.com/huggingface/lerobot/issues/2465
open
[ "bug", "policies", "training" ]
2025-11-18T03:46:28Z
2025-12-03T16:13:56Z
null
Lilgeneric
huggingface/lerobot
2,464
Questions about Pi0.5 Model Training Details and High Level Planning Implementation
Hello, while studying the Pi0.5 model, I have two questions regarding the model implementation that I would like to ask you: 1、The paper mentions that the model adopts two-stage pre-training and designs a comprehensive loss function. However, when checking the compute_loss part in the open-source code, it is found that currently only the action loss is calculated, and the loss related to the VLM (Vision-Language Model) in the pre-training stage is not reflected. I would like to confirm whether this part is implemented elsewhere in the code or if there are other design considerations? 2、The ablation experiments in the paper show that the jointly trained Pi0.5 performs excellently in explicit and implicit High Level planning, even better than GPT4 and manual upper-level planning. However, from the open-source model code, the implementation part related to the High Level planning step has not been found for the time being. I would like to know how this part of the function is reflected in the code? Looking forward to your reply, thank you!
https://github.com/huggingface/lerobot/issues/2464
open
[ "question", "training" ]
2025-11-18T01:27:59Z
2025-11-20T10:45:34Z
null
Ginldaj
vllm-project/vllm
28,876
[CI Failure]: should test_cumem.py use spawn or fork in cuda?
### Name of failing test tests/basic_correctness/test_cumem.py ### Basic information - [ ] Flaky test - [x] Can reproduce locally - [ ] Caused by external libraries (e.g. bug in `transformers`) ### 🧪 Describe the failing test The test only fails locally for me when I use vllm main branch and on the CI of my PR, error is caused by cuda tests using `fork` instead of `spawn` I think, in the CI, there is a line that's trying for force spawn: https://github.com/vllm-project/vllm/blob/f2b8e1c5510cf3621dc4b910f0eba5289d9fee88/.buildkite/test-pipeline.yaml#L99-L100, but looks like it's not effective. I looked at the function that decides to use fork or spawn: https://github.com/vllm-project/vllm/blob/f8b19c0ffd65f7f6f01a0da4a39b6890f5db40cb/tests/utils.py#L1027 and I don't think it looks like the flag `VLLM_WORKER_MULTIPROC_METHOD`. Although the issue doesn't repro in the main vllm CI. Wondering how do we fix this? ``` FAILED basic_correctness/test_cumem.py::test_python_error - RuntimeError: Cannot re-initialize CUDA in forked subprocess. To use CUDA with multiprocessing, you must use the 'spawn' start method FAILED basic_correctness/test_cumem.py::test_basic_cumem - RuntimeError: Cannot re-initialize CUDA in forked subprocess. To use CUDA with multiprocessing, you must use the 'spawn' start method FAILED basic_correctness/test_cumem.py::test_cumem_with_cudagraph - RuntimeError: Cannot re-initialize CUDA in forked subprocess. To use CUDA with multiprocessing, you must use the 'spawn' start method FAILED basic_correctness/test_cumem.py::test_end_to_end[hmellor/tiny-random-LlamaForCausalLM] - RuntimeError: Cannot re-initialize CUDA in forked subprocess. To use CUDA with multiprocessing, you must use the 'spawn' start method FAILED basic_correctness/test_cumem.py::test_end_to_end[facebook/opt-125m] - RuntimeError: Cannot re-initialize CUDA in forked subprocess. To use CUDA with multiprocessing, you must use the 'spawn' start method FAILED basic_correctness/test_cumem.py::test_deep_sleep - RuntimeError: Cannot re-initialize CUDA in forked subprocess. To use CUDA with multiprocessing, you must use the 'spawn' start method FAILED basic_correctness/test_cumem.py::test_deep_sleep_async - RuntimeError: Cannot re-initialize CUDA in forked subprocess. To use CUDA with multiprocessing, you must use the 'spawn' start method ``` ### 📝 History of failing test https://buildkite.com/vllm/ci/builds/39127/steps/canvas?jid=019a84f5-0fbf-46f3-859f-42c02a2d3de1 ### CC List. _No response_
https://github.com/vllm-project/vllm/issues/28876
open
[ "ci-failure" ]
2025-11-17T18:58:08Z
2025-11-17T20:59:14Z
1
jerryzh168
vllm-project/vllm
28,868
[Bug]: When compiling with ranges, we should pass the range information to Inductor
### Your current environment main ### 🐛 Describe the bug Might be more of a feature request. Context is that https://github.com/vllm-project/vllm/pull/24248 adds a new compile ranges API, where a user can specify which ranges to compile on. We should tell Inductor how to constrain the compilation on the symints of the compile ranges ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/28868
open
[ "bug", "torch.compile" ]
2025-11-17T15:41:50Z
2026-01-05T23:37:12Z
1
zou3519
vllm-project/vllm
28,866
[Usage]: When is going to be the next release?
Hi everyone, Thank you for developing such a great tool! I was wondering when the next release is scheduled. I’m interested in running Gemma3-text type architecture GGUF quantized models with VLLM. Are there any alternatives to do this with the latest release (v0.11.0)? I also noticed that you merged this PR with the working solution on October 9: https://github.com/vllm-project/vllm/pull/26189
https://github.com/vllm-project/vllm/issues/28866
open
[ "usage" ]
2025-11-17T15:24:47Z
2025-11-19T10:51:47Z
1
Invalid-coder
huggingface/transformers
42,241
How to use padding with Mistral?
I'm trying to understand how to use Mistral with `batch_size` > 1. One aspect of this is setting `padding="longest"` in, e.g., `MistralCommonTokenizer.encode()`. But I'm getting `TypeError: 'set' object is not callable` when I try this. Example: ```python import torch from transformers import MistralForCausalLM, MistralCommonTokenizer tokenizer = MistralCommonTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.3") model = MistralForCausalLM.from_pretrained( "mistralai/Mistral-7B-Instruct-v0.3", dtype=torch.bfloat16, attn_implementation="sdpa", device_map="auto", ) messages = [ "You are a pirate chatbot who always responds in pirate speak!", "Who are you?", ] model_inputs = tokenizer.encode(messages, return_tensors="pt", padding="longest").to( model.device ) ``` Output: ``` --------------------------------------------------------------------------- TypeError Traceback (most recent call last) Cell In[1], line 17 5 model = MistralForCausalLM.from_pretrained( 6 "mistralai/Mistral-7B-Instruct-v0.3", 7 dtype=torch.bfloat16, 8 attn_implementation="sdpa", 9 device_map="auto", 10 ) 12 messages = [ 13 "You are a pirate chatbot who always responds in pirate speak!", 14 "Who are you?", 15 ] ---> 17 model_inputs = tokenizer.encode(messages, return_tensors="pt", padding="longest").to( 18 model.device 19 ) 21 generated_ids = model.generate(model_inputs, max_new_tokens=100, do_sample=True) 22 tokenizer.batch_decode(generated_ids)[0] File ~/ad_hoc_analysis/src/asr_and_summarization/.venv/lib/python3.13/site-packages/transformers/tokenization_mistral_common.py:407, in MistralCommonTokenizer.encode(self, text, text_pair, add_special_tokens, padding, truncation, max_length, stride, pad_to_multiple_of, padding_side, return_tensors, verbose, **kwargs) 404 if text_pair: 405 raise ValueError("`MistralCommonTokenizer.encode` does not support `text_pair`.") --> 407 padding_strategy, truncation_strategy, max_length, _ = self._get_padding_truncation_strategies( 408 padding=padding, 409 truncation=truncation, 410 max_length=max_length, 411 pad_to_multiple_of=pad_to_multiple_of, 412 verbose=verbose, 413 ) 415 encoded_inputs = self._encode_plus( 416 text, 417 add_special_tokens=add_special_tokens, (...) 429 verbose=verbose, 430 ) 432 return encoded_inputs["input_ids"] File ~/ad_hoc_analysis/src/asr_and_summarization/.venv/lib/python3.13/site-packages/transformers/tokenization_mistral_common.py:1034, in MistralCommonTokenizer._get_padding_truncation_strategies(self, padding, truncation, max_length, pad_to_multiple_of, verbose, **kwargs) 1031 max_length = self.model_max_length 1033 # Test if we have a padding token -> 1034 if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.pad_token is None or self.pad_token_id < 0): 1035 raise ValueError( 1036 "Asking to pad but the tokenizer does not have a padding token. " 1037 "Please select a token to use as `pad_token` `(tokenizer.pad_token = tokenizer.eos_token e.g.)` " 1038 "or add a new pad token via `tokenizer.add_special_tokens({'pad_token': '[PAD]'})`." 1039 ) 1041 # Check that we will truncate to a multiple of pad_to_multiple_of if both are provided File ~/ad_hoc_analysis/src/asr_and_summarization/.venv/lib/python3.13/site-packages/transformers/tokenization_mistral_common.py:334, in MistralCommonTokenizer.pad_token(self) 329 @property 330 def pad_token(self) -> str: 331 """ 332 String associated to the padding token in the vocabulary. 333 """ --> 334 return self.convert_ids_to_tokens(self.pad_token_id) File ~/ad_hoc_analysis/src/asr_and_summarization/.venv/lib/python3.13/site-packages/transformers/tokenization_mistral_common.py:548, in MistralCommonTokenizer.convert_ids_to_tokens(self, ids, skip_special_tokens) 546 tokens: list[str] = [] 547 for token_id in ids: --> 548 if self._is_control_token(token_id) and skip_special_tokens: 549 continue 550 tokens.append(self.tokenizer.instruct_tokenizer.tokenizer.id_to_piece(token_id)) File ~/ad_hoc_analysis/src/asr_and_summarization/.venv/lib/python3.13/site-packages/transformers/tokenization_mistral_common.py:513, in MistralCommonTokenizer._is_control_token(self, token_id) 511 def _is_control_token(self, token_id: int) -> bool: 512 if self._tokenizer_type == MistralTokenizerType.spm: --> 513 return token_id in self.tokenizer.instruct_tokenizer.tokenizer._control_tokens() 514 elif self._tokenizer_type == MistralTokenizerType.tekken: 515 return token_id < self.tokenizer.instruct_tokenizer.tokenizer.num_special_tokens TypeError: 'set' object is not callable ``` Env: ``` - `transformers` version: 4.57.1
https://github.com/huggingface/transformers/issues/42241
closed
[]
2025-11-17T12:54:21Z
2025-11-19T06:11:44Z
null
TopCoder2K
huggingface/chat-ui
1,986
HI i would like to use default_headers={ "X-HF-Bill-To": "org-name" } in my chatui local deployment how i can??
Hi, So i want to bill my Inference usage to my organization and like to pass default_headers={ "X-HF-Bill-To": "org-name" } parameter how i can do that??
https://github.com/huggingface/chat-ui/issues/1986
open
[ "support" ]
2025-11-17T08:33:41Z
2025-11-17T08:33:41Z
null
aditya-oss-prog
huggingface/diffusers
12,672
How to set pipe "requires_grad=true"?
I have set the variable and the model "requires_grad=true" with the following: ` pipe.transformer.requires_grad = True pipe.vae.requires_grad = True` `prev_sample = prev_sample.detach().requires_grad_(True)` but the "requires_grad" of result by the pipe is still not true: `image_tar = pipe.vae.decode(prev_sample, return_dict=False)[0]` "image_tar" still can not requires_grad, so how to set pipe "requires_grad=true"?(all the operation is during inference stage.)
https://github.com/huggingface/diffusers/issues/12672
closed
[]
2025-11-17T03:36:43Z
2025-11-20T12:19:20Z
null
micklexqg
huggingface/diffusers
12,669
Flux1-Dev inference with single file ComfyUI/SD-Forge Safetensors
Is it possible to run inference with diffusers using a single-file safetensors created for ComfyUI/SD-Forge? It looks like FluxPipeline.from_single_file() might be intended for this purpose, but I'm getting the following errors: ``` import torch from diffusers import FluxPipeline pipe = FluxPipeline.from_single_file("./flux1-dev-fp8.safetensors", torch_dtype=torch.float8_e4m3fn, use_safetensors=True) ``` ``` Traceback (most recent call last): File "/home/user/flux/imgen.py", line 9, in <module> pipe = FluxPipeline.from_single_file("./flux1-dev-fp8.safetensors", torch_dtype=torch.float8_e4m3fn, use_safetensors=True) File "/home/user/.local/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "/home/user/.local/lib/python3.13/site-packages/diffusers/loaders/single_file.py", line 509, in from_single_file loaded_sub_model = load_single_file_sub_model( library_name=library_name, ...<11 lines>... **kwargs, ) File "/home/user/.local/lib/python3.13/site-packages/diffusers/loaders/single_file.py", line 127, in load_single_file_sub_model loaded_sub_model = create_diffusers_t5_model_from_checkpoint( class_obj, ...<4 lines>... local_files_only=local_files_only, ) File "/home/user/.local/lib/python3.13/site-packages/diffusers/loaders/single_file_utils.py", line 2156, in create_diffusers_t5_model_from_checkpoint model.load_state_dict(diffusers_format_checkpoint) ~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/user/.local/lib/python3.13/site-packages/torch/nn/modules/module.py", line 2641, in load_state_dict raise RuntimeError( ...<3 lines>... ) RuntimeError: Error(s) in loading state_dict for T5EncoderModel: Missing key(s) in state_dict: "encoder.embed_tokens.weight". ``` I checked the safetensors file and the T5 encoder is present. However, it is named differently, which confuses diffusers.
https://github.com/huggingface/diffusers/issues/12669
open
[]
2025-11-16T11:57:48Z
2025-12-03T16:53:58Z
12
ddpasa
huggingface/ai-deadlines
41
How to indicate ARR deadlines
Right now the yaml format assumes conferences with locations and dates, but ACL ARR has rolling deadlines not tied to a physical conference. We are largely operating around these deadlines. How can we incorporate these into this system?
https://github.com/huggingface/ai-deadlines/issues/41
open
[]
2025-11-15T00:26:33Z
2025-11-15T00:26:33Z
null
morrisalp
huggingface/diffusers
12,662
question on stable_audio_transformer.py
Execuse me, I am leaning the code of `class StableAudioDiTModel` , I do not know what is the argument ` global_states_input_dim` used to? It seems that it is a must component that should be packed before the hidden_states sequence. and its default dim seems larger then the transformer inner_dim. What is that componenet means? If it is used to take in additional conditions, that seems can be done in the encoder outside. and compared with the concatenate, I think it may be better to repeat condition embedding to the sequence length and concat on hidden_dim. And what is the ` sample_size: int = 1024,` parameter used in the model creation? it seems not used during `forward` call The func doc of `class StableAudioDiTModel:forward`, it said ``` encoder_attention_mask (`torch.Tensor` of shape `(batch_size, sequence_len)`, *optional*):```. why the shape of encoder_attention_mask is batch_size X sequence_len instead of batch_size X encoder_sequence_len to be identical with the shape of the input `encoder_hidden_states` and why thee return value of this `forward` is the direct `(hidden_states,)` but not `(hidden_states * attention_mask, )`? about the `class StableAudioDiTModel forward`, what is the shape of parameters `rotary_embedding` and `timestep`? why the global_embedding is concated before the hidden_states? I think hidden_states is what we want to generated during DiT pipeline. while encoder_hidden_states is the condition signal, so global_embedding should be used to en-rich the encoder_hidden_states. and the action of concate the global_embedding before the input hidden_states sequence will change the input seq_length, according to[ [1]](https://github.com/Stability-AI/stable-audio-tools/blob/main/docs/conditioning.md#input-concatenation), the concatenation should be done in the feature_dim direction, is it? It seems using normal LayerNorm layer instead of adaLN layer?
https://github.com/huggingface/diffusers/issues/12662
open
[]
2025-11-14T09:26:01Z
2025-11-25T08:53:39Z
1
JohnHerry
vllm-project/vllm
28,717
[Usage]: Errors running vLLM docker in a closed environment with gpt-oss-120b on RTX 6000 Pro
### Your current environment Can't get vLLM to start with the below configuration. Seems to have issues loading in the model .safetensors. Any ideas on what could be causing it? vllm version: 0.11.1 CPU: Intel Xeon w7-2595X GPU: NVIDIA RTX PRO 6000 Blackwell Max-Q Workstation Edition Model: https://huggingface.co/openai/gpt-oss-120b/tree/main Command: docker run --rm --name vllm --gpus=all --runtime=nvidia -p 8000:8000 -e HF_HUB_OFFLINE=1 --ipc=host -v opt/models/cache/:/root/.cache/huggingface/hub vllm/vllm-openai:latest --model openai/gpt-oss-120b Also tried: docker run --rm --name vllm --gpus=all --runtime=nvidia -p 8000:8000 -e HF_HUB_OFFLINE=1 --ipc=host -v opt/models/cache/:/root/.cache/huggingface/hub vllm/vllm-openai:latest --model openai/gpt-oss-120b with the same output. Output: INFO 11-12 06:23:18 [__init__.py:216] Automatically detected platform cuda. (APIServer pid=1) INFO 11-12 06:23:21 [api_server.py:1839] vLLM API server version 0.11.0 (APIServer pid=1) INFO 11-12 06:23:21 [utils.py:233] non-default args: {'model': 'openai/gpt-oss-120b'} (APIServer pid=1) INFO 11-12 06:23:21 [arg_utils.py:504] HF_HUB_OFFLINE is True, replace model_id [openai/gpt-oss-120b] to model_path [/root/.cache/huggingface/hub/models--openai--gpt-oss-120b/snapshots/b5c939de8f754692c1647ca79fbf85e8c1e70f8a] (APIServer pid=1) `torch_dtype` is deprecated! Use `dtype` instead! (APIServer pid=1) INFO 11-12 06:23:26 [model.py:547] Resolved architecture: GptOssForCausalLM (APIServer pid=1) ERROR 11-12 06:23:26 [config.py:278] Error retrieving safetensors: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/root/.cache/huggingface/hub/models--openai--gpt-oss-120b/snapshots/b5c939de8f754692c1647ca79fbf85e8c1e70f8a'. Use `repo_type` argument if needed., retrying 1 of 2 (APIServer pid=1) ERROR 11-12 06:23:28 [config.py:276] Error retrieving safetensors: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/root/.cache/huggingface/hub/models--openai--gpt-oss-120b/snapshots/b5c939de8f754692c1647ca79fbf85e8c1e70f8a'. Use `repo_type` argument if needed. (APIServer pid=1) INFO 11-12 06:23:28 [model.py:1730] Downcasting torch.float32 to torch.bfloat16. (APIServer pid=1) INFO 11-12 06:23:28 [model.py:1510] Using max model len 131072 (APIServer pid=1) INFO 11-12 06:23:29 [scheduler.py:205] Chunked prefill is enabled with max_num_batched_tokens=8192. (APIServer pid=1) INFO 11-12 06:23:29 [config.py:271] Overriding max cuda graph capture size to 992 for performance. INFO 11-12 06:23:31 [__init__.py:216] Automatically detected platform cuda. (EngineCore_DP0 pid=308) INFO 11-12 06:23:33 [core.py:644] Waiting for init message from front-end. (EngineCore_DP0 pid=308) INFO 11-12 06:23:33 [core.py:77] Initializing a V1 LLM engine (v0.11.0) with config: model='/root/.cache/huggingface/hub/models--openai--gpt-oss-120b/snapshots/b5c939de8f754692c1647ca79fbf85e8c1e70f8a', speculative_config=None, tokenizer='/root/.cache/huggingface/hub/models--openai--gpt-oss-120b/snapshots/b5c939de8f754692c1647ca79fbf85e8c1e70f8a', skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.bfloat16, max_seq_len=131072, download_dir=None, load_format=auto, tensor_parallel_size=1, pipeline_parallel_size=1, data_parallel_size=1, disable_custom_all_reduce=False, quantization=mxfp4, enforce_eager=False, kv_cache_dtype=auto, device_config=cuda, structured_outputs_config=StructuredOutputsConfig(backend='auto', disable_fallback=False, disable_any_whitespace=False, disable_additional_properties=False, reasoning_parser='openai_gptoss'), observability_config=ObservabilityConfig(show_hidden_metrics_for_version=None, otlp_traces_endpoint=None, collect_detailed_traces=None), seed=0, served_model_name=/root/.cache/huggingface/hub/models--openai--gpt-oss-120b/snapshots/b5c939de8f754692c1647ca79fbf85e8c1e70f8a, enable_prefix_caching=True, chunked_prefill_enabled=True, pooler_config=None, compilation_config={"level":3,"debug_dump_path":"","cache_dir":"","backend":"","custom_ops":[],"splitting_ops":["vllm.unified_attention","vllm.unified_attention_with_output","vllm.mamba_mixer2","vllm.mamba_mixer","vllm.short_conv","vllm.linear_attention","vllm.plamo2_mamba_mixer","vllm.gdn_attention","vllm.sparse_attn_indexer"],"use_inductor":true,"compile_sizes":[],"inductor_compile_config":{"enable_auto_functionalized_v2":false},"inductor_passes":{},"cudagraph_mode":[2,1],"use_cudagraph":true,"cudagraph_num_of_warmups":1,"cudagraph_capture_sizes":[992,976,960,944,928,912,896,880,864,848,832,816,800,784,768,752,736,720,704,688,672,656,640,624,608,592,576,560,544,528,512,496,480,464,448,432,416,400,384,368,352,336,320,304,288,272,256,248,240,232,224,216,208,200,192,184,176,168,160,152,144,136,128,120,112,104,96,88,80,72,64,56,48
https://github.com/vllm-project/vllm/issues/28717
open
[ "usage" ]
2025-11-14T08:49:48Z
2025-11-20T15:45:21Z
3
antonkarlsson1
huggingface/trl
4,525
How to modify the advantage computation in GRPOTrainer
I’m looking to customize the advantage computation used in the DAPO algorithm. Do I need to subclass the full GRPOTrainer to do this, or is it sufficient to overwrite the logic in _generate_and_score_completions, since that method appears to handle the advantage calculation?
https://github.com/huggingface/trl/issues/4525
open
[ "❓ question", "🏋 GRPO" ]
2025-11-14T03:48:17Z
2025-11-14T11:37:18Z
null
Tuziking
huggingface/transformers
42,200
Request of rewriting implementation of prediction_step in trainer.py
### System Info Any system. Because it's a problem coming from source code. ### Who can help? @SunMarc ### Information - [ ] The official example scripts - [x] My own modified scripts ### Tasks - [ ] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...) - [x] My own task or dataset (give details below) ### Reproduction Hi, i am talking about an issue that was reported 5 years ago but still exists in 2025, specifically, 13th Nov, 2025. I quote one of the issues that was discussed before, ignored by sgugger. Please find the link below https://discuss.huggingface.co/t/cuda-out-of-memory-when-using-trainer-with-compute-metrics/2941 When i was about to fine tune a LLM today, i ran into the same issue but i got saved by one folk's solution provided in this discussion. How to reproduce (you should have a GPU, no quantization, just full fine tuning): 1. Find a random decoder-only text2text LLM, let's say Qwen3 0.6B. 2. Prepare a train dataset (>0 rows) and eval dataset (>850 rows). 3. Set eval_on_start = True, either TrainingArguments or SFTConfig could work. 4. Implement your own compute_metrics BUT DON'T implement preprocess_logits_for_metrics. 5. start training (don't need deepspeed or accelerate, just trainer.train()) What would happen? First it would go through the evaluation dataset because i set eval_on_start=True, the model would go really fast originally but then it would go extremely slow. Finally, you would get an error that says numpy is trying to allocate a ridiculously big array to memory. <img width="1567" height="986" alt="Image" src="https://github.com/user-attachments/assets/e1885324-fb09-48b6-8bfd-d36306c2a156" /> One of the folk who seems to be inspired by example code provided the implementation of preprocess_logits_for_metrics, which solved problem i encountered perfectly. The evaluation run is done within 2 mins. Why it would happen? I briefly go over the source code of evaluation_loop and i located prediction_step. prediction_step says it would return a tuple of three optional torch.Tensor (loss, logits, label). <img width="719" height="68" alt="Image" src="https://github.com/user-attachments/assets/537032b1-9371-4852-bed8-8f31cd6a0437" /> But most of the time, the returned logits is a tuple. Why? if you look at the the function that processes logits before logits is returned: <img width="535" height="140" alt="Image" src="https://github.com/user-attachments/assets/d6f7f3b1-6c2a-4298-b4c2-f0ab85fa88cf" /> This function would receive all kinds of "tensors". The type of "tensors" could be list, tuple, Mapping or torch.Tensor. Does it change the variable, called "tensors", from other data types to torch.Tensor? No. type(tensors)(........) would preserve the original type of tensors. It means if the variable "tensors" (i hate this variable name because it is misleading and confusing) is a tuple, after this function, it's still a tuple!!!!! It's a recursive function btw. I would love doing recursion in programming competition, but not in huggingface codebase!!! It also implies a fact that the input of nested_detach could be complexly nested, like ([],()) So this function doesn't guarantee the logits is a torch.Tensor. Nor does the implementation of prediction_step before nested_detach was called in prediction_step <img width="702" height="759" alt="Image" src="https://github.com/user-attachments/assets/451982b4-648b-4876-a2b7-c9d748899fd1" /> So, the logits is not always a torch.Tensor, which is contradictory to what the type hint says, what did developers do? They developed preprocess_logits_for_metrics. So that user could fix it ON THEIR OWN IMPLEMENTATION. (preprocess_logits_for_metrics is called within evaluation_loop to clean the mess, specifically, logits, returned by prediction_step()) <img width="803" height="772" alt="Image" src="https://github.com/user-attachments/assets/c7494018-e282-4577-b824-3db9c9e57609" /> It's such a lazy fix. Why a regular user is expected to implement their own preprocess_logits_f or_metrics, to deal with a poorly-designed prediction_step? It has been 5 years since the person who reported it......... If a user-defined compute_metrics is not provided to Trainer or SFTTrainer, the prediction_step would return (loss, none, none), which skips the whole problem and this is why users said the issue of "slow evaluation" is gone when they don't provide compute_metrics. I would like to make a Pull Request to fix it but i don't have enough time and energy to do this massive amount of work. A temporary fix is to let users know when they need to make their own compute_metrics, they also have to implement preprocess_logits_for_metrics. Different models would have different styles of implementations but for text2text decoder only LLM. <img width="687" height="78" alt="Image" src="https://github.com/user-attachments/assets/125fffe3-d8cc-44c7-9a96-35a11500d975" /> (Another thing is that the variable called
https://github.com/huggingface/transformers/issues/42200
open
[ "Good Second Issue", "bug" ]
2025-11-14T00:13:40Z
2025-12-18T14:29:32Z
3
Yacklin
huggingface/transformers
42,197
Attempt to access socket despite HF_HUB_OFFLINE = 1 if cache warmed outside current process
### System Info - `transformers` version: 4.57.1 - Platform: Linux-6.6.84.1-microsoft-standard-WSL2-x86_64-with-glibc2.35 - Python version: 3.13.0 - Huggingface_hub version: 0.36.0 - Safetensors version: 0.6.2 - Accelerate version: not installed - Accelerate config: not found - DeepSpeed version: not installed - PyTorch version (accelerator?): 2.9.1+cpu (NA) - Tensorflow version (GPU?): not installed (NA) - Flax version (CPU?/GPU?/TPU?): not installed (NA) - Jax version: not installed - JaxLib version: not installed - Using distributed or parallel set-up in script?: No ### Who can help? @ydshieh I have created a reproducible example of the issue I mentioned in https://github.com/huggingface/transformers/issues/41311#issuecomment-3508674325. ### Information - [ ] The official example scripts - [x] My own modified scripts ### Tasks - [ ] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...) - [x] My own task or dataset (give details below) ### Reproduction Reproducible example: https://github.com/fr1ll/HF_HUB_OFFLINE Warming the cache in a subprocess, then disabling sockets, then loading the same model should work. However, it fails with an attempt to access a socket and then "Can't load" errors. The script named `subprocess-warm_then_offline-load.py` reproduces this error. Interestingly, warming the cache in process, then disabling sockets, then loading the same model works. This is reproduced in `inprocess-warm_then_offline-load.py` in the repo above. ### Expected behavior When a model has already been loaded into the cache ("warm cache"), if `HF_OFFLINE_MODE` = `"1"`, a Transformers pipeline should be able to load the model without accessing any network sockets.
https://github.com/huggingface/transformers/issues/42197
closed
[ "Good Second Issue", "bug" ]
2025-11-13T21:38:29Z
2025-11-24T09:33:54Z
6
fr1ll
vllm-project/vllm
28,646
[Feature][P2]: Implement CI Build Time and Size Guards
### 🚀 The feature, motivation and pitch ### Description Once we optimize the Docker build, we need to prevent regressions. Create CI checks that fail if build time exceeds thresholds or if image size grows beyond acceptable limits. Also set up monitoring dashboards. ### What You'll Do 1. Create Python scripts to check image metrics: - `check-image-size.py` (extend existing wheel size checker) - `check-build-time.py` - `check-image-layers.py` 2. Add these checks to CI pipeline after image build 3. Set appropriate thresholds (configurable) 4. Create Buildkite annotations for warnings 5. Set up CloudWatch dashboard for metrics (optional) ### Deliverables - [ ] Python scripts for checking metrics - [ ] Integration into test-template-ci.j2 - [ ] Configurable thresholds via environment variables - [ ] Documentation on how to adjust thresholds - [ ] CloudWatch dashboard (optional) ### Alternatives _No response_ ### Additional context _No response_ ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/28646
open
[ "feature request", "ci/build" ]
2025-11-13T12:50:34Z
2025-11-13T18:55:29Z
0
rzabarazesh
huggingface/diffusers
12,650
Question about the `# Copied from` system
Hi team! 👋 While working on improving docstrings and type hints across scheduler files (issue #9567), I've noticed the `# Copied from` pattern used extensively throughout the codebase. Examples: - Functions like `betas_for_alpha_bar` are duplicated across multiple schedulers - Output classes like `DDPMSchedulerOutput` are copied with name replacements (e.g., DDPM->EulerDiscrete) My question: What's the rationale behind this duplication system instead of: 1. Using a shared utils.py or common.py file for common functions 2. Using class inheritance for similar Output classes I understand there might be good architectural reasons (module independence, API stability, avoiding circular dependencies, etc.), but this isn't documented anywhere that I could find. Suggested action: Regardless of the answer, I think we should either: - Option A: Refactor to use inheritance/shared utilities (if the current system is legacy) - Option B: Document this design decision in: &nbsp; - A CONTRIBUTING.md or architecture doc &nbsp; - Comments in the utils/check_copies.py script itself &nbsp; - Another README in the diffusers directory This would help future contributors (like me! 😅) understand why this pattern exists and how to work with it properly when improving documentation. What do you think? Thanks for maintaining such a great library! 🚀
https://github.com/huggingface/diffusers/issues/12650
open
[]
2025-11-13T11:53:22Z
2025-12-21T22:44:03Z
3
delmalih
huggingface/transformers
42,179
Add TileLang Kernel Support
### Feature request I would like to propose adding support for TileLang kernel in the transformers library. TileLang is a modular approach for writing attention kernels that could provide flexibility and performance benefits. github link: https://github.com/tile-ai/tilelang - Add TileLang as an optional attention backend - Provide configuration options similar to existing attention mechanisms - Ensure compatibility with existing model architectures - Add proper multi-GPU support and synchronization ### Motivation - Enhanced Modularity TileLang offers a more modular approach to writing attention kernels, making it easier for researchers and developers to modify and optimize the implementation for specific use cases. - Performance Comparison Integrating TileLang would allow users to benchmark its performance directly against existing attention implementations, such as Flex Attention and Flash Attention. This would foster a better understanding of how different kernels can impact model performance and efficiency. - Community Engagement Supporting TileLang in the Transformers library would attract a broader community of developers interested in optimizing transformer models, thus enhancing collaboration and innovation. - Flexibility TileLang's architecture is designed for ease of modification, allowing users to experiment with and refine attention mechanisms more effectively. ### Your contribution I've experimented with TileLang kernel on transformers models and found it works well in single-GPU scenarios. However, when enabling multi-GPU inference using `device_map='auto'`, I encounter NaN tensors. This may be related to tensor synchronization issues in distributed settings. I'm willing to help with testing and potentially contributing to the implementation once the multi-GPU synchronization issue is understood and resolved. I also have 3 questions: 1. Is there any existing plan or roadmap for TileLang integration? 2. Are there specific guidelines for adding new attention backends? 3. What would be the recommended approach for handling multi-GPU synchronization in custom kernels?
https://github.com/huggingface/transformers/issues/42179
open
[ "Feature request" ]
2025-11-13T11:38:33Z
2025-11-13T11:38:33Z
0
crownz248
huggingface/tokenizers
1,885
Feature request: Characters delimiter argument
I wish to develop a k-mer-character-based BPE tokenizer using your beautiful Rust package, for genomic applications. Unfortunately, it doesn't seem to support defining a characters delimiter. As I see it, it is a pretty straightforward change, instead of iterating a word by character, first split it by the delimiter and then iterate. Also, when merges are computed, in the string representation the character delimiter should also be considered. In that way, a multi-character word splitting could have been made feasible. Right now I am using a modified Python version of the BPE tokenizer made by the genius [Yikai-Liao](https://github.com/Yikai-Liao/efficient_bpe/blob/main/ebpe_v2.py), however it would be nice to see that happening in Rust as well, and natively supported by huggingface. Unfortunately, I am still novice in working with Rust, otherwise I would make a pull request with the suggested changes. Is it something that can be worked out in the future? Or is there a way to do this with the current implementation? Thank you!
https://github.com/huggingface/tokenizers/issues/1885
open
[]
2025-11-13T10:40:29Z
2025-11-28T07:51:07Z
1
VasLem
vllm-project/vllm
28,629
[Usage]: TPOT per request information was not collected by vllm bench serve
### Your current environment ```text The output of `python collect_env.py` Collecting environment information... ============================== System Info ============================== OS : Ubuntu 24.04.2 LTS (x86_64) GCC version : (Ubuntu 13.3.0-6ubuntu2~24.04) 13.3.0 Clang version : Could not collect CMake version : version 4.1.0 Libc version : glibc-2.39 ============================== PyTorch Info ============================== PyTorch version : 2.8.0+xpu Is debug build : False CUDA used to build PyTorch : None ROCM used to build PyTorch : N/A ============================== Python Environment ============================== Python version : 3.12.3 (main, Aug 14 2025, 17:47:21) [GCC 13.3.0] (64-bit runtime) Python platform : Linux-6.14.0-1006-intel-x86_64-with-glibc2.39 ============================== CUDA / GPU Info ============================== Is CUDA available : False CUDA runtime version : No CUDA CUDA_MODULE_LOADING set to : N/A GPU models and configuration : No CUDA Nvidia driver version : No CUDA cuDNN version : No CUDA HIP runtime version : N/A MIOpen runtime version : N/A Is XNNPACK available : True ============================== CPU Info ============================== Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 46 bits physical, 57 bits virtual Byte Order: Little Endian CPU(s): 32 On-line CPU(s) list: 0-31 Vendor ID: GenuineIntel BIOS Vendor ID: Intel(R) Corporation Model name: Intel(R) Xeon(R) w5-3435X BIOS Model name: Intel(R) Xeon(R) w5-3435X CPU @ 3.1GHz BIOS CPU family: 179 CPU family: 6 Model: 143 Thread(s) per core: 2 Core(s) per socket: 16 Socket(s): 1 Stepping: 8 CPU(s) scaling MHz: 45% CPU max MHz: 4700.0000 CPU min MHz: 800.0000 BogoMIPS: 6192.00 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf tsc_known_freq pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb cat_l3 cat_l2 cdp_l3 intel_ppin cdp_l2 ssbd mba ibrs ibpb stibp ibrs_enhanced tpr_shadow flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid cqm rdt_a avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb intel_pt avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local split_lock_detect user_shstk avx_vnni avx512_bf16 wbnoinvd dtherm ida arat pln pts hwp hwp_act_window hwp_epp hwp_pkg_req vnmi avx512vbmi umip pku ospke waitpkg avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid bus_lock_detect cldemote movdiri movdir64b enqcmd fsrm md_clear serialize tsxldtrk pconfig arch_lbr ibt amx_bf16 avx512_fp16 amx_tile amx_int8 flush_l1d arch_capabilities Virtualization: VT-x L1d cache: 768 KiB (16 instances) L1i cache: 512 KiB (16 instances) L2 cache: 32 MiB (16 instances) L3 cache: 45 MiB (1 instance) NUMA node(s): 1 NUMA node0 CPU(s): 0-31 Vulnerability Gather data sampling: Not affected Vulnerability Ghostwrite: Not affected Vulnerability Indirect target selection: Not affected Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Not affected Vulnerability Reg file data sampling: Not affected Vulnerability Retbleed: Not affected Vulnerability Spec rstack overflow: Not affected Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and
https://github.com/vllm-project/vllm/issues/28629
open
[ "usage" ]
2025-11-13T09:20:19Z
2025-11-13T09:20:19Z
0
jlwang1996
vllm-project/vllm
28,626
[Bug]:Qwen3-VL-32B-AWQ model memory usage: 8k context limit with 40GB VRAM?
### Your current environment <details> <summary>The output of <code>python collect_env.py</code></summary> ```text Your output of `python collect_env.py` here ``` </details> ### 🐛 Describe the bug Running models on the latest stable vLLM release: https://huggingface.co/QuantTrio/Qwen3-VL-32B-Instruct-AWQ The model size is 20GB, and my GPU has 40GB VRAM total. Using parameter: --gpu-memory-utilization 0.9 Why am I only getting around 8k max context length? Do VL models really hog that much VRAM? ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/28626
open
[ "bug" ]
2025-11-13T08:00:20Z
2025-11-17T07:08:47Z
3
maxin9966
vllm-project/vllm
28,622
[Bug]: Can we able to benchmark Quantized MOE models Either W8A8 or W8A16 ?
### Your current environment <details> <summary>The output of <code>python collect_env.py</code></summary> ```text Collecting environment information... ============================== System Info ============================== OS : Ubuntu 24.04.2 LTS (x86_64) GCC version : (Ubuntu 13.3.0-6ubuntu2~24.04) 13.3.0 Clang version : Could not collect CMake version : version 3.22.1 Libc version : glibc-2.39 ============================== PyTorch Info ============================== PyTorch version : 2.8.0+cu128 Is debug build : False CUDA used to build PyTorch : 12.8 ROCM used to build PyTorch : N/A ============================== Python Environment ============================== Python version : 3.10.18 (main, Jun 4 2025, 08:56:00) [GCC 13.3.0] (64-bit runtime) Python platform : Linux-6.14.0-33-generic-x86_64-with-glibc2.39 ============================== CUDA / GPU Info ============================== Is CUDA available : True CUDA runtime version : 12.0.140 CUDA_MODULE_LOADING set to : LAZY GPU models and configuration : GPU 0: NVIDIA RTX 6000 Ada Generation Nvidia driver version : 575.57.08 cuDNN version : Probably one of the following: /usr/lib/x86_64-linux-gnu/libcudnn.so.9.11.0 /usr/lib/x86_64-linux-gnu/libcudnn_adv.so.9.11.0 /usr/lib/x86_64-linux-gnu/libcudnn_cnn.so.9.11.0 /usr/lib/x86_64-linux-gnu/libcudnn_engines_precompiled.so.9.11.0 /usr/lib/x86_64-linux-gnu/libcudnn_engines_runtime_compiled.so.9.11.0 /usr/lib/x86_64-linux-gnu/libcudnn_graph.so.9.11.0 /usr/lib/x86_64-linux-gnu/libcudnn_heuristic.so.9.11.0 /usr/lib/x86_64-linux-gnu/libcudnn_ops.so.9.11.0 HIP runtime version : N/A MIOpen runtime version : N/A Is XNNPACK available : True ============================== CPU Info ============================== Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 52 bits physical, 57 bits virtual Byte Order: Little Endian CPU(s): 52 On-line CPU(s) list: 0-51 Vendor ID: GenuineIntel Model name: Intel(R) Xeon(R) w7-2595X CPU family: 6 Model: 143 Thread(s) per core: 2 Core(s) per socket: 26 Socket(s): 1 Stepping: 8 CPU(s) scaling MHz: 21% CPU max MHz: 4800.0000 CPU min MHz: 800.0000 BogoMIPS: 5616.00 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf tsc_known_freq pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb cat_l3 cat_l2 cdp_l3 intel_ppin cdp_l2 ssbd mba ibrs ibpb stibp ibrs_enhanced tpr_shadow flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid cqm rdt_a avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb intel_pt avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local split_lock_detect user_shstk avx_vnni avx512_bf16 wbnoinvd dtherm ida arat pln pts hwp hwp_act_window hwp_epp hwp_pkg_req vnmi avx512vbmi umip pku ospke waitpkg avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid bus_lock_detect cldemote movdiri movdir64b enqcmd fsrm md_clear serialize tsxldtrk pconfig arch_lbr ibt amx_bf16 avx512_fp16 amx_tile amx_int8 flush_l1d arch_capabilities Virtualization: VT-x L1d cache: 1.2 MiB (26 instances) L1i cache: 832 KiB (26 instances) L2 cache: 52 MiB (26 instances) L3 cache: 48.8 MiB (1 instance) NUMA node(s): 1 NUMA node0 CPU(s): 0-51 Vulnerability Gather data sampling: Not affected Vulnerability Ghostwrite: Not affected Vulnerability Indirect target selection: Not affected Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Not af
https://github.com/vllm-project/vllm/issues/28622
open
[ "bug" ]
2025-11-13T07:26:56Z
2025-11-13T07:27:06Z
0
logesh13
vllm-project/vllm
28,610
[Usage]: Does 0.11.0 suport tree attenton with eagle?
### Your current environment Does 0.11.0 suport tree attenton with eagle? Do I need to enable it manually? ### How would you like to use vllm I want to run inference of a [specific model](put link here). I don't know how to integrate it with vllm. ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/28610
open
[ "usage" ]
2025-11-13T03:35:02Z
2025-12-03T17:08:16Z
1
wincle
huggingface/datasets
7,864
add_column and add_item erroneously(?) require new_fingerprint parameter
### Describe the bug Contradicting their documentation (which doesn't mention the parameter at all), both Dataset.add_column and Dataset.add_item require a new_fingerprint string. This parameter is passed directly to the dataset constructor, which has the fingerprint parameter listed as optional; is there any reason it shouldn't be optional in these methods as well? ### Steps to reproduce the bug Reproduction steps: 1. Look at the function signature for add_column: https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L6078 2. Repeat for add_item: https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L6336 ### Expected behavior add_column and add_item should either set the fingerprint parameter to optional or include it in their docstrings ### Environment info Not environment-dependent
https://github.com/huggingface/datasets/issues/7864
open
[]
2025-11-13T02:56:49Z
2025-12-07T14:41:40Z
2
echthesia
vllm-project/vllm
28,566
[Usage]: pd disagg scenario , I discover in the decoder , also has the prefill operation, is it normal ?
### Your current environment when num_computed_tokens is less than num_prompt_tokens, it will enter prefill operation <img width="633" height="149" alt="Image" src="https://github.com/user-attachments/assets/bab96187-37c8-4ea2-ba68-9f52dda07f6b" /> and i found, num_computed_tokens is possible less than num_prompt_tokens, because num_prompt_tokens is len(block_ids) * self.block_size, event num_prompt_tokens is just equal to num_prompt_tokens, it do num_computed_tokens -= 1, why ? this cause num_computed_tokens is never equal to num_prompt_tokens <img width="980" height="762" alt="Image" src="https://github.com/user-attachments/assets/81eb6f4f-f0db-45f8-8934-64bd8ea21988" /> ### How would you like to use vllm I want to run inference of a [specific model](put link here). I don't know how to integrate it with vllm. ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
https://github.com/vllm-project/vllm/issues/28566
open
[ "usage" ]
2025-11-12T16:18:53Z
2025-11-12T16:18:53Z
0
yangshanjun