Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ggml: aarch64: implement mmla kernels for q8_0_q8_0, q4_0_q8_0 and q4_1_q8_1 quantized gemm #4966

Merged
merged 5 commits into from
Feb 11, 2024

llama.cpp: add MATMUL_INT8 capability to system_info

d8f132d
Select commit
Loading
Failed to load commit list.
Sign in for the full log view
Merged

ggml: aarch64: implement mmla kernels for q8_0_q8_0, q4_0_q8_0 and q4_1_q8_1 quantized gemm #4966

llama.cpp: add MATMUL_INT8 capability to system_info
d8f132d
Select commit
Loading
Failed to load commit list.

Annotations

1 warning

The logs for this run have expired and are no longer available.