-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
src: cpu: aarch64: Re-enable matmul static quantisation through ACL. #2308
Conversation
88cf5e0
to
5cec04d
Compare
Hi @mgouicem, it's the same as before but more restrictive with Neoverse-N1 until the failing tests are fixed. |
Hi @renato-arantes I see that you added a merge commit, please rebase your branch instead as we try to keep a linear history. |
ce7f10b
to
66d9ea4
Compare
@renato-arantes, please resolve conflicts. |
66d9ea4
to
b322d6d
Compare
b322d6d
to
38d8255
Compare
arm_compute::QuantizationInfo(*src_scale, -src_zero_point, true)); | ||
acl_obj.wei_tensor.info()->set_quantization_info( | ||
arm_compute::QuantizationInfo(*wei_scale, -wei_zero_point, true)); | ||
// for efficiency reasons, OneDNN saves the inverse of the destination |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
It's oneDNN
, not OneDNN
. Fixup is here.
Description
This is to re-enable
matmul
static quantization operations through ACL. Currently, the supported data type combinations ares8:s8:s8
andu8:s8:u8
. Neoverse-N1 is skipped until failing tests are fixed.Previous PR: #2198
General
make test
andmake test_benchdnn_*
) pass locally for each commit?