multi-lora batching
#14249
-
Does vllm support batching of prompts with different lora-adapters? Is there a more detailed example code? |
Beta Was this translation helpful? Give feedback.
Answered by
jeejeelee
Mar 5, 2025
Replies: 1 comment 1 reply
-
vLLM support this featrue
|
Beta Was this translation helpful? Give feedback.
1 reply
Answer selected by
Diffizle
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
vLLM support this featrue
multilora_inference.py
show this feature, differentlora_id
to represent different LoRAs.