Replies: 1 comment
-
Here's the list of currently supported ops with the QNN EP. More info directly from Qualcomm here I ran into similar issues with some of my model ops not being supported and just had to let it fall back to the CPU. It should do this by default I believe unless you set the |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hello everyone,
I’m experimenting with running OpenAI’s Whisper-Medium model via ONNX Runtime on a Windows laptop with a Snapdragon X Elite NPU.
• Tiny/Base models run successfully on the NPU (QNN Execution Provider).
• Medium/Large models always fail “Failed to finalize QNN graph. Error code: 1002” and fall back to GPU (AzureExecutionProvider).
I’ve confirmed:
It looks like some ONNX operators used by Whisper-Medium aren’t yet implemented in the QNN plugin. I’m not a deep-learning engineer—just a power user—so I’m not sure which operators are missing or how to work around it.
My questions:
I can share logs, minimal repro code and ONNX export steps if helpful. Thank you for any guidance or pointers!
—
Environment:
Beta Was this translation helpful? Give feedback.
All reactions