Arm backend: Add 16A8W support and test for mul operation#13785
Arm backend: Add 16A8W support and test for mul operation#13785Ninja91 wants to merge 1 commit intogh/Ninja91/4/basefrom
Conversation
Add 16A8W quantization support and test for the mul operation in ExecutorTorch ARM backend. This follows the pattern established for linear operations, extending int16 support to mul operations. Changes: - Add INT16 dtype validation support in op_mul.py - Add test_mul_tensor_16a8w_tosa_INT test function - Enable test_mul.py in test targets configuration The 16A8W configuration uses 16-bit activations with 8-bit weights, enabling higher precision for activations while maintaining weight efficiency. Differential Revision: [D80510628](https://our.internmc.facebook.com/intern/diff/D80510628/) [ghstack-poisoned]
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/13785
Note: Links to docs will display an error until the docs builds have been completed. ❌ 3 New FailuresAs of commit 11571f9 with merge base 9053089 ( NEW FAILURES - The following jobs have failed:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
This PR needs a
|
|
Looks like this PR hasn't been updated in a while so we're going to go ahead and mark this as |
|
Looks like this PR hasn't been updated in a while so we're going to go ahead and mark this as |
|
Is this still needed? |
|
We think this is done and can be closed, please reopen of you don't agree. |
Stack from ghstack (oldest at bottom):
Add 16A8W quantization support and test for the mul operation in ExecutorTorch ARM backend.
This follows the pattern established for linear operations, extending int16 support to mul operations.
Changes:
The 16A8W configuration uses 16-bit activations with 8-bit weights, enabling higher precision for activations while maintaining weight efficiency.
Differential Revision: D80510628
cc @freddan80 @per @zingo @oscarandersson8218 @digantdesai