You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@tianleiwu thank you,
I also have a problem. I run a single-core single instance on a 92.8 cpu. The computing power required for the model is 654.66 GFlops. The theoretical reasoning time is 654.66/92.8 = 7s. However, the actual test takes 15s. Is the gap normal? Why?
Describe the issue
I have 128 cpu cores,when I use onnx to inference, OMP_NUM_THREADS=1 numactl -c 1 python xxx and python xxx got same performance,why?
To reproduce
start = time.perf_counter()
output = session.run()
print(time.perf_counter()-start)
Urgency
No response
Platform
Linux
OS Version
openeuler
ONNX Runtime Installation
Released Package
ONNX Runtime Version or Commit ID
1.19.2
ONNX Runtime API
Python
Architecture
ARM64
Execution Provider
Default CPU
Execution Provider Library Version
No response
Model File
No response
Is this a quantized model?
Yes
The text was updated successfully, but these errors were encountered: