Skip to content

Commit 0922bc4

Browse files
committed
fix full graph
Signed-off-by: ganyi <ygan@amd.com>
1 parent fd2edfb commit 0922bc4

File tree

6 files changed

+319
-459
lines changed

6 files changed

+319
-459
lines changed

vllm/config/scheduler.py

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -137,10 +137,6 @@ class SchedulerConfig:
137137
structured outputs, speculative decoding, and pipeline parallelism.
138138
"""
139139

140-
split_prefill_from_chunk: bool = False
141-
"""Whether to split the prefill request into pure prefill and chunked
142-
prefill in a single batch."""
143-
144140
def compute_hash(self) -> str:
145141
"""
146142
WARNING: Whenever a new field is added to this config,

vllm/platforms/rocm.py

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -390,11 +390,6 @@ def check_and_update_config(cls, vllm_config: "VllmConfig") -> None:
390390
):
391391
compilation_config.custom_ops.append("+rms_norm")
392392

393-
if envs.VLLM_ROCM_USE_AITER and envs.VLLM_ROCM_USE_AITER_MHA:
394-
# enable the request reorder if we are using AITER MHA
395-
# for calculation
396-
vllm_config.scheduler_config.split_prefill_from_chunk = True
397-
398393
@classmethod
399394
def verify_model_arch(cls, model_arch: str) -> None:
400395
if model_arch in _ROCM_UNSUPPORTED_MODELS:

0 commit comments

Comments
 (0)