Skip to content

Commit e77713c

Browse files
committed
[misc] Remove useless patch
Signed-off-by: wangxiyuan <[email protected]>
1 parent e985432 commit e77713c

File tree

3 files changed

+1
-50
lines changed

3 files changed

+1
-50
lines changed

vllm_ascend/patch/__init__.py

Lines changed: 1 addition & 23 deletions
Original file line numberDiff line numberDiff line change
@@ -104,29 +104,7 @@
104104
# Future Plan:
105105
# Remove this patch when vllm merged them.
106106
#
107-
# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
108-
# 1. `vllm.v1.sample.sampler.Sampler.gather_logprobs`
109-
# Why:
110-
# We need to patch gather_logprobs to make sure call batched_count_greater_than
111-
# with backend=current_platform.simple_compile_backend
112-
# How:
113-
# Patch gather_logprobs call new batched_count_greater_than
114-
# Related PR (if no, explain why):
115-
# - https://github.com/vllm-project/vllm/pull/21591
116-
# Future Plan:
117-
# Revert it when vLLM merge #21591 and release new version
118-
# ** File: worker/patch_logits.py **
119-
# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
120-
# 1. `vllm._custom_ops.apply_repetition_penalties`
121-
# Why:
122-
# apply_repetition_penalties in vLLM use tensor.is_cuda to check if tensor is on cuda. But the value is always True
123-
# on ascend, thus we need to patch apply_repetition_penalties.
124-
# How:
125-
# Remove the related cuda check in apply_repetition_penalties.
126-
# Related PR (if no, explain why):
127-
# - this is a bug by Ascend only. It can' be fixed in vLLM.
128-
# Future Plan:
129-
# Fix this bug in torch-npu, bump torch-npu version and remove this patch.
107+
# ** File: worker/patch_roberta.py **
130108
# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
131109
# 1. `vllm.model_executor.models.roberta.RobertaEmbedding.forward`
132110
# Why:

vllm_ascend/patch/worker/__init__.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,6 @@
2323
# isort: off
2424
import vllm_ascend.patch.platform.patch_sched_yield # noqa
2525
import vllm_ascend.patch.worker.patch_distributed # noqa
26-
import vllm_ascend.patch.worker.patch_logits # noqa
2726
import vllm_ascend.patch.worker.patch_roberta # noqa
2827
import vllm_ascend.patch.worker.patch_weight_loader # noqa
2928
import vllm_ascend.patch.worker.patch_multimodal_merge # noqa

vllm_ascend/patch/worker/patch_logits.py

Lines changed: 0 additions & 26 deletions
This file was deleted.

0 commit comments

Comments
 (0)