-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add Qualcomm mobile SoC native backend for GGML #771
Comments
It seems done(although lack of implementation of other GGML OPs using QNN API ) and now it's PR stage. Implementation of GGML_OP_MUL_MAT using QNN API and GGML_OP_ADD using QNN API and GGML_OP_MUL using QNN API could be found in ggml-qnn.cpp. other implementation of GGML OPs using QNN API is very similar to GGML_OP_ADD / GGML_OP_MUL / GGML_OP_MULMAT in ggml-qnn.cpp. hope community could help to do it in upstream whiserp.cpp after the PR is accepted by upstream whisper.cpp so I'd like to close this opening issue accordingly. |
@zhouwg Thanks for you work :) Has the PR been merged? I cannot find QNN in the llama.cpp nor whisper.cpp supported backend list |
we should/might study something about Qualcomm's dedicated/proprietary machine learning acceleration library accordingly.
Opinions about template based GEMM kernels llama.cpp#3965
Add GPU support to ggml llama.cpp#915
Adding Native Support of SYCL for Intel GPUs llama.cpp#4749
SYCL backend support Multi-card llama.cpp#5282
The text was updated successfully, but these errors were encountered: