ColossalAI VS Vllm Benchmark #5513
Replies: 2 comments
-
|
Beta Was this translation helpful? Give feedback.
-
Hey @zzb610 , (I'll use English to reply so that anyone could participate into the discussion) |
Beta Was this translation helpful? Give feedback.
-
我在 A100 40G 上使用 ColossalAI 的 colossal-infer 分支 (main 分支跑不通) 的代码进行了对 llama-7b 模型进行了推理性能的测试
https://github.com/hpcaitech/ColossalAI/tree/feature/colossal-infer/colossalai/inference
CollossalAI 和 vllm 的版本分别是
colossalai 21e1e36
vllm 0.3.0
得到的结果如下
请问各位大佬, 造成 ColossalAI 在 bs > 32 推理性能超过 vllm 的原因是什么 ?
flash_decoding_attention ? KVCachaManager ? RequestHandler
Beta Was this translation helpful? Give feedback.
All reactions