dheeena's picture
Add files using upload-large-folder tool
9496ebf verified
# coding=utf-8
# Copyright 2025 The HuggingFace Inc. team
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import threading
from abc import ABC, abstractmethod
from collections import deque
from ...utils.metrics import attach_tracer, traced
from .cache import PagedAttentionCache
from .requests import RequestState, RequestStatus
class Scheduler(ABC):
"""
Abstract base class for scheduling requests in the continuous batch processor. Schedulers manage the lifecycle of
requests from when they are added to the waiting queue to when they are scheduled for processing. Different
schedulers implement different strategies for prioritizing and batching requests.
"""
def __init__(self, cache: PagedAttentionCache, retain_cache_on_finish: bool = False):
self.active_requests: dict[str, RequestState] = {}
self.waiting_requests: dict[str, RequestState] = {}
self.waiting_requests_order: deque[str] = deque()
self.cache = cache
self.retain_cache_on_finish = retain_cache_on_finish
self._cancellation_lock = threading.Lock()
self._requests_to_cancel: set[str] = set()
@traced
def add_waiting_request(self, state: RequestState):
"""Adds a request to the waiting list."""
if self.retain_cache_on_finish and state.request_id in self.active_requests:
old_state = self.active_requests.pop(state.request_id)
state.prompt_ids = state.prompt_ids[len(old_state.full_prompt_ids) :] # XXX: check for indexing error?
state.allocated_blocks = old_state.allocated_blocks
state.position_offset = old_state.position_offset
self.waiting_requests[state.request_id] = state
self.waiting_requests_order.append(state.request_id)
@abstractmethod
def schedule_batch(self, token_budget: int) -> list[RequestState]:
"""Schedules requests for the next batch based on available token budget. This method selects which requests
should be processed in the current batch, considering the token budget and the scheduler's prioritization rules.
The token_budget is the maximum number of tokens that can be processed in this batch."""
pass
@traced
def has_pending_requests(self) -> bool:
"""Checks if there are requests ready to be processed."""
return len(self.active_requests) or len(self.waiting_requests)
@traced
def finish_request(self, request_id: str, evict_from_cache: bool = True):
"""Completes processing of a request and optionally frees its allocated cache blocks. This method is called
when a request has finished generation or encountered an error.
"""
if evict_from_cache:
self.cache.free_blocks(request_id)
if request_id in self.active_requests:
del self.active_requests[request_id]
@traced
def get_active_request_static_outputs(self, request_id: str) -> list[int]:
"""Gets generated tokens for an active request."""
if request_id in self.active_requests:
return self.active_requests[request_id].static_outputs
return []
@traced
def set_request_cancellation(self, request_id: str):
"""Marks a request for cancellation."""
with self._cancellation_lock:
self._requests_to_cancel.add(request_id)
@traced
def clear_cancelled_requests(self):
"""Remove all cancelled requests from active and waiting queues."""
with self._cancellation_lock:
for request_id in self._requests_to_cancel:
if request_id in self.active_requests:
del self.active_requests[request_id]
if request_id in self.waiting_requests:
del self.waiting_requests[request_id]
if request_id in self.waiting_requests_order:
self.waiting_requests_order.remove(request_id)
self.cache.free_blocks(request_id)
self._requests_to_cancel = set()
@traced
def request_is_cancelled(self, request_id: str) -> bool:
"""Checks if a request has been cancelled or removed."""
return request_id in self._requests_to_cancel or (
request_id not in self.active_requests and request_id not in self.waiting_requests
)
@traced
def _allocate_blocks_if_needed(self, state: RequestState, len_next_tokens: int) -> bool:
"""Allocate additional cache blocks for a request if the currently allocated blocks are insufficient to
accommodate the next tokens. It calculates how many blocks are needed based on the request's current
cache occupancy and the number of tokens to be processed. The allocation itself is done by the CacheAllocator
objects. Returns a boolean indicating if the allocation was successful or not.
"""
# 1. we check that the occupancy is less than the requested length
# 2. we allocate enough blocks to cover the requested length
current_len = state.current_len()
occupancy = state.allocated_blocks * self.cache.block_size - current_len
if occupancy < len_next_tokens or state.allocated_blocks == 0:
blocks_needed = ((len_next_tokens - occupancy + 1) // self.cache.block_size) + 1
allocated = self.cache.allocate_blocks(blocks_needed, state.request_id)
if allocated is None:
return False
state.allocated_blocks += allocated
return True
@traced(span_name="prepare_request")
def _prepare_request_for_processing(
self, state: RequestState, token_budget: int, request_ids_to_remove_from_waiting: set[str]
):
"""Prepares a request for processing in the current batch."""
request_tokens = (
state.remaining_prompt_ids if state.status == RequestStatus.SPLIT_PENDING_REMAINDER else state.prompt_ids
)
if len(request_tokens) < token_budget:
# Can process the entire prompt/remainder
if state.status == RequestStatus.PENDING:
self.active_requests[state.request_id] = state
state.status = RequestStatus.PREFILLING
request_ids_to_remove_from_waiting.add(state.request_id)
elif state.status == RequestStatus.SPLIT_PENDING_REMAINDER:
state.status = RequestStatus.PREFILLING
state.prompt_ids = state.remaining_prompt_ids
state.remaining_prompt_ids = []
else:
# Need to split the request
if state.status == RequestStatus.PENDING:
self.active_requests[state.request_id] = state
state.status = RequestStatus.PREFILLING_SPLIT
request_ids_to_remove_from_waiting.add(state.request_id)
elif state.status == RequestStatus.SPLIT_PENDING_REMAINDER:
state.status = RequestStatus.PREFILLING_SPLIT
state.remaining_prompt_ids = request_tokens[token_budget:]
state.prompt_ids = request_tokens[:token_budget]
@attach_tracer()
class FIFOScheduler(Scheduler):
"""This scheduler processes requests in the order they arrive, meaning decoding requests has priority over
prefilling requests. Additionally, it includes a safety margin mechanism to prevent cache exhaustion. By default,
when 80% of the cache is full, new requests will not be scheduled to prioritize decoding active requests."""
def __init__(self, cache: PagedAttentionCache, retain_cache_on_finish: bool = False, safety_margin: float = 0.2):
"""Initializes the FIFO scheduler. The safety margin is the percentage of free blocks under which we stop
scheduling new prefill requests, so safety_margin = 0.1 means that when there is less than 10% of free blocks,
or equivalently when more than 90% of blocks are already allocated, we stop scheduling new prefill requests.
"""
super().__init__(cache, retain_cache_on_finish)
self.safety_margin = safety_margin
@traced
def schedule_batch(self, token_budget: int) -> list[RequestState]:
priority_states: list[RequestState] = []
second_priority_states: list[RequestState] = []
scheduled_requests = []
for state in self.active_requests.values():
if state.status == RequestStatus.DECODING:
priority_states.append(state)
if state.status in [RequestStatus.SPLIT_PENDING_REMAINDER, RequestStatus.PREFILLING_SPLIT]:
second_priority_states.append(state)
# Add waiting requests to second priority
for req_id in self.waiting_requests_order:
second_priority_states.append(self.waiting_requests[req_id])
candidates = priority_states + second_priority_states
request_ids_to_remove_from_waiting = set()
safety_margins = self.safety_margin * self.cache.num_blocks
for state in candidates:
# If we are out the safety margin, we only accept decoding requests or the first prefill request
num_free_blocks = self.cache.get_num_free_blocks()
outside_safety_margin = num_free_blocks < safety_margins
if outside_safety_margin and scheduled_requests and state.status != RequestStatus.DECODING:
break
self._prepare_request_for_processing(state, token_budget, request_ids_to_remove_from_waiting)
request_len = len(state.prompt_ids)
if not self._allocate_blocks_if_needed(
state, len(state.prompt_ids)
): # don't schedule if we can't allocate blocks
if len(self.cache._free_blocks) == 0:
break
continue
@traced
def _add_to_scheduled_requests(state: RequestState):
scheduled_requests.append(state)
_add_to_scheduled_requests(state)
token_budget -= request_len
@traced
def _remove_from_waiting_requests(state: RequestState):
req_id = state.request_id
if req_id in self.waiting_requests:
del self.waiting_requests[req_id]
request_ids_to_remove_from_waiting.add(req_id)
_remove_from_waiting_requests(state)
if token_budget == 0:
break
self.waiting_requests_order = deque(
[req_id for req_id in self.waiting_requests_order if req_id not in request_ids_to_remove_from_waiting]
)
return scheduled_requests
# FIXME: prioritize adding from waiting reqs before scheduling `RequestStatus.DECODING` when cache space allows it
@attach_tracer()
class PrefillFirstScheduler(Scheduler):
"""Scheduler that prioritizes split prefill requests over decoding requests. This scheduler ensures that split
prefill requests (which are continuations of partially processed prompts) are completed before processing new
decoding requests."""
@traced
def schedule_batch(self, token_budget: int) -> list[RequestState]:
priority_states: list[RequestState] = []
second_priority_states: list[RequestState] = []
scheduled_requests = []
for state in self.active_requests.values():
# XXX: when cache is full, state can stay on `PREFILLING_SPLIT` so we need to take those into account
if state.status in [RequestStatus.PREFILLING_SPLIT, RequestStatus.SPLIT_PENDING_REMAINDER]:
priority_states.append(state)
elif state.status == RequestStatus.DECODING:
second_priority_states.append(state)
for req_id in self.waiting_requests_order:
second_priority_states.append(self.waiting_requests[req_id])
candidates = priority_states + second_priority_states
request_ids_to_remove_from_waiting = set()
for state in candidates:
self._prepare_request_for_processing(state, token_budget, request_ids_to_remove_from_waiting)
request_len = len(state.prompt_ids)
if not self._allocate_blocks_if_needed(
state, len(state.prompt_ids)
): # don't schedule if we can't allocate blocks
if len(self.cache._free_blocks) == 0:
break
continue
@traced
def _add_to_scheduled_requests(state: RequestState):
scheduled_requests.append(state)
_add_to_scheduled_requests(state)
token_budget -= request_len
@traced
def _remove_from_waiting_requests(state: RequestState):
req_id = state.request_id
if req_id in self.waiting_requests:
del self.waiting_requests[req_id]
request_ids_to_remove_from_waiting.add(req_id)
_remove_from_waiting_requests(state)
if token_budget == 0:
break
self.waiting_requests_order = deque(
[req_id for req_id in self.waiting_requests_order if req_id not in request_ids_to_remove_from_waiting]
)
return scheduled_requests
SCHEDULER_MAPPING = {
"fifo": FIFOScheduler,
"prefill_first": PrefillFirstScheduler,
}