-
Describe the issueWondering if there's a performance difference between using "pip install onnxruntime" versus building onnxruntime on device (using instructions like https://onnxruntime.ai/docs/build/inferencing.html#linux). I use CPUExecutionProvider, on Ubuntu 22.04, on a computer that uses RK3588 (which has ARMv8.2 CPUs). I currently use "pip install onnxruntime", and things work perfectly fine, but am wondering if more performance can be squeezed out. To reproduceNot a bug, just a general performance question. UrgencyNot urgent, just a general performance question. PlatformLinux OS VersionUbuntu 22.04 ONNX Runtime InstallationReleased Package ONNX Runtime Version or Commit ID1.15.1 ONNX Runtime APIPython ArchitectureARM64 Execution ProviderDefault CPU Execution Provider Library VersionNo response Model FileNo response Is this a quantized model?No |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 2 replies
-
though the base arch used is armv8-a, onnxruntime does support dynamic dispatch of the arch specific kernels during runtime. For example, there are mmla, fp16 and bf16 kernels supported in python wheel. So, I doubt if anything is missing in the wheel and can be extracted in source builds. btw, do you have any particular kernel in mind? |
Beta Was this translation helpful? Give feedback.
though the base arch used is armv8-a, onnxruntime does support dynamic dispatch of the arch specific kernels during runtime. For example, there are mmla, fp16 and bf16 kernels supported in python wheel. So, I doubt if anything is missing in the wheel and can be extracted in source builds. btw, do you have any particular kernel in mind?