-
Notifications
You must be signed in to change notification settings - Fork 3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
add op_types_to_quantize to get_qnn_qdq_config #23458
base: main
Are you sure you want to change the base?
Conversation
onnxruntime/python/tools/quantization/execution_providers/qnn/quant_config.py
Outdated
Show resolved
Hide resolved
Looks like there are some linter errors. Should be able to fix them by running |
Never thought it should be written in this way.. |
Hi @xieofxie, could you please try merging the latest changes from main? That should get the CI working. |
Sure, merged @adrianlizarraga Just resolved conflict, please take a look, thanks! |
Description
add op_types_to_quantize to get_qnn_qdq_config so support only quantize part of it.
In my understanding, if ops are not quantized, they will use fp16 to run if supported?
https://docs.qualcomm.com/bundle/publicresource/topics/80-63442-50/SupportedOps.html
Motivation and Context
See scenario in issue in Olive microsoft/Olive#1552