You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Theodoree
changed the title
Misc. bug: runing failure on Adreno devices using Vulkan for large batch size
Misc. bug: Vulkan Error: Device Lost on Adreno GPUs with Large Batch Sizes
Mar 2, 2025
Theodoree
changed the title
Misc. bug: Vulkan Error: Device Lost on Adreno GPUs with Large Batch Sizes
Misc. bug: Device Lost on Adreno GPUs with Large Batch Sizes
Mar 2, 2025
Theodoree
changed the title
Misc. bug: Device Lost on Adreno GPUs with Large Batch Sizes
Misc. bug: vulkan on Adreno GPU
Mar 2, 2025
Name and Version
llama-cli --version
ggml_vulkan: Found 1 Vulkan devices:
ggml_vulkan: 0 = Adreno (TM) 732 (Qualcomm Technologies Inc. Adreno Vulkan Driver) | uma: 1 | fp16: 1 | warp size: 64 | shared memory: 32768 | matrix cores: none
version: 4798 (1782cdf)
built with Android (12896553, +pgo, -bolt, +lto, -mlgo, based on r530567c) clang version 19.0.0 (https://android.googlesource.com/toolchain/llvm-project 97a699bf4812a18fb657c2779f5296a4ab2694d2) for x86_64-apple-darwin24.3.0
Operating systems
Linux
Which llama.cpp modules do you know to be affected?
llama-bench
Command line
Problem description & steps to reproduce
model: lb-reranker-0.5B-v1.0-Q4_0.gguf
Problem Description
When running
llama-bench
, I encountered the following error when thebatch-size
exceeds 32:Expected Behavior
The benchmark should run successfully with any
batch-size
, including values larger than 32.Actual Behavior
The program crashes with the
DeviceLost
error when thebatch-size
is greater than 32.Environment Information
llama-bench
batch-size
: Greater than 32vk::DeviceLostError
Build Script Used
First Bad Commit
No response
Relevant log output
The text was updated successfully, but these errors were encountered: