Skip to content

Commit e251425

Browse files
gcanlinwangyu31577
authored andcommitted
[Clean] Remove the redundant decoding payloads logic (vllm-project#404)
Signed-off-by: gcanlin <canlinguosdu@gmail.com> Signed-off-by: wangyu31577 <wangyu31577@hundsun.com>
1 parent 1c5b276 commit e251425

File tree

1 file changed

+0
-3
lines changed

1 file changed

+0
-3
lines changed

vllm_omni/worker/gpu_model_runner.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -739,9 +739,6 @@ def _preprocess(
739739
intermediate_tensors: IntermediateTensors | None = None,
740740
):
741741
"""Align with v0.12 preprocess and omni's additional information handling."""
742-
# Decode payload first, ensure request state has prompt_embeds / additional_information
743-
self._decode_and_store_request_payloads(scheduler_output)
744-
745742
num_scheduled_tokens = scheduler_output.total_num_scheduled_tokens
746743
is_first_rank = get_pp_group().is_first_rank
747744
is_encoder_decoder = self.model_config.is_encoder_decoder

0 commit comments

Comments
 (0)