Skip to content

Commit

Permalink
doc
Browse files Browse the repository at this point in the history
Signed-off-by: Cody Yu <[email protected]>
  • Loading branch information
comaniac committed Feb 13, 2025
1 parent 568be95 commit 82567a3
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions vllm/v1/core/scheduler.py
Original file line number Diff line number Diff line change
Expand Up @@ -302,7 +302,6 @@ def schedule(self) -> "SchedulerOutput":

# Get the longest common prefix among all requests in the running queue.
# This can be potentially used for cascade attention.
# FIXME: This is not correct.
num_common_prefix_blocks = 0
if self.running:
any_request = self.running[0]
Expand Down Expand Up @@ -463,7 +462,7 @@ def update_from_output(
req_id = request.request_id
num_tokens_scheduled = num_scheduled_tokens.get(req_id, 0)
if num_tokens_scheduled == 0:
# The request was not scheduled in this batch.
# The request was not scheduled in this step.
new_running.append(request)
continue

Expand Down Expand Up @@ -610,6 +609,7 @@ def has_unfinished_requests(self) -> bool:
return self.get_num_unfinished_requests() > 0

def get_num_unscheduled_requests(self) -> int:
"""Number of requests that are not being processed by the executor."""
return self.get_num_unfinished_requests() - len(self.scheduled_req_ids)

def reset_prefix_cache(self) -> bool:
Expand Down

0 comments on commit 82567a3

Please sign in to comment.