Skip to content

Commit 3709433

Browse files
committed
fix
Signed-off-by: Woosuk Kwon <[email protected]>
1 parent 2985f65 commit 3709433

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

vllm/v1/worker/gpu/spec_decode/eagle.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -139,7 +139,7 @@ def generate_draft(
139139
num_tokens_across_dp: torch.Tensor | None,
140140
) -> None:
141141
pos = self.input_buffers.positions[:num_reqs]
142-
query_start_loc = self.input_buffers.query_start_loc[: num_reqs + 1]
142+
query_start_loc = self.input_buffers.query_start_loc.gpu[: num_reqs + 1]
143143
for step in range(1, self.num_speculative_steps):
144144
# Run the eagle model.
145145
last_hidden_states, hidden_states = self.run_model(

0 commit comments

Comments
 (0)