UPSTREAM PR #19433: Add a build target to generate ROCm artifacts using ROCm 7.2#1160
UPSTREAM PR #19433: Add a build target to generate ROCm artifacts using ROCm 7.2#1160
Conversation
|
No meaningful performance changes were detected across 115630 analyzed functions in the following binaries: build.bin.llama-cvector-generator, build.bin.llama-tts, build.bin.libllama.so, build.bin.libmtmd.so, build.bin.libggml-cpu.so, build.bin.libggml.so, build.bin.libggml-base.so, build.bin.llama-bench, build.bin.llama-gemma3-cli, build.bin.llama-gguf-split, build.bin.llama-llava-cli, build.bin.llama-minicpmv-cli, build.bin.llama-quantize, build.bin.llama-tokenize, build.bin.llama-qwen2vl-cli. 🔎 Full breakdown: Loci Inspector. |
|
@loci-dev how many functions were modified in this pr and how many files were edited or added ? |
|
Based on the repository analysis: PR StatisticsFiles Changed
Functions AnalyzedAccording to the performance review, 115,630 functions were analyzed across all binaries, with no meaningful performance changes detected in any of them. This is a substantial PR focused on:
The fact that 115K+ functions were analyzed but showed no meaningful performance regressions is actually a positive signal - it indicates that despite the extensive changes across 118 files, the modifications maintained performance stability. The changes appear to be primarily focused on correctness, feature additions, and code quality improvements rather than introducing performance-impacting modifications. |
ef7afbe to
d4c3480
Compare
8aeb553 to
d961293
Compare
|
No meaningful performance changes were detected across 115001 analyzed functions in the following binaries: build.bin.llama-cvector-generator, build.bin.llama-tts, build.bin.libggml-cpu.so, build.bin.libmtmd.so, build.bin.libllama.so, build.bin.libggml-base.so, build.bin.llama-bench, build.bin.llama-quantize, build.bin.llama-tokenize, build.bin.llama-gguf-split, build.bin.libggml.so, build.bin.llama-qwen2vl-cli, build.bin.llama-gemma3-cli, build.bin.llama-llava-cli, build.bin.llama-minicpmv-cli. 🔎 Full breakdown: Loci Inspector. |
073bd79 to
823244c
Compare
This builds the following targets: * gfx1151 * gfx1150 * gfx1200 * gfx1201 * gfx1100 * gfx1101 * gfx908 * gfx90a * gfx942
d961293 to
4a1f236
Compare
|
No meaningful performance changes were detected across 111508 analyzed functions in the following binaries: build.bin.llama-cvector-generator, build.bin.libllama.so, build.bin.libmtmd.so, build.bin.llama-tts, build.bin.llama-tokenize, build.bin.llama-gemma3-cli, build.bin.llama-gguf-split, build.bin.llama-llava-cli, build.bin.llama-minicpmv-cli, build.bin.llama-quantize, build.bin.llama-qwen2vl-cli, build.bin.llama-bench, build.bin.libggml-base.so, build.bin.libggml-cpu.so, build.bin.libggml.so. 🔎 Full breakdown: Loci Inspector. |
10f8f26 to
a6ecec6
Compare
Note
Source pull request: ggml-org/llama.cpp#19433
This builds the following targets:
Notably; it doesn't include any CDNA targets, and thus avoids ggml-org/llama.cpp#19269