Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions backends/cadence/aot/BUCK
Original file line number Diff line number Diff line change
Expand Up @@ -45,6 +45,7 @@ fbcode_target(_kind = runtime.python_library,
":utils",
"//caffe2:torch",
"//executorch/backends/cadence/aot/quantizer:fusion_pass",
"//executorch/backends/cadence/aot/quantizer/passes:fuse_ops",
"//executorch/backends/cadence/aot/quantizer:quantizer",
"//executorch/backends/transforms:decompose_sdpa",
"//executorch/backends/transforms:remove_clone_ops",
Expand Down
14 changes: 10 additions & 4 deletions backends/cadence/aot/compiler.py
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,7 @@
print_memory_planning_info,
)
from executorch.backends.cadence.aot.quantizer.fusion_pass import QuantFusion
from executorch.backends.cadence.aot.quantizer.passes.fuse_ops import FuseQATConvBN
from executorch.backends.cadence.aot.quantizer.quantizer import (
CadenceDefaultQuantizer,
CadenceQuantizer,
Expand All @@ -37,9 +38,10 @@
ExecutorchBackendConfig,
ExecutorchProgramManager,
)
from executorch.exir.pass_manager import PassManager
from executorch.exir.passes import ToOutVarPass
from executorch.exir.passes.sym_shape_eval_pass import ConstraintBasedSymShapeEvalPass
from executorch.exir.program._program import _transform, to_edge
from executorch.exir.program._program import to_edge
from torch.export.exported_program import ExportedProgram
from torchao.quantization.pt2e.quantize_pt2e import convert_pt2e

Expand Down Expand Up @@ -162,13 +164,17 @@ def apply_pre_edge_transform_passes(
which will instantiate a default quantizer for you if needed.
Returns an ExportedProgram with the fused model.
"""
# Get patterns and apply fusion of dq -> op -> q to qop
# pyre-ignore[16]: no attribute
patterns = [q.pattern for q in quantizer.quantizers]
fused_program = _transform(converted_program, QuantFusion(patterns))
PassManager(
[
FuseQATConvBN(converted_program),
QuantFusion(patterns),
]
)(converted_program.graph_module)

# Apply torch ops passes (e.g., ReplaceMulTensorWithMulAndFullOpsPass)
fused_program = apply_torch_ops_passes(fused_program)
fused_program = apply_torch_ops_passes(converted_program)

return fused_program

Expand Down
16 changes: 16 additions & 0 deletions backends/cadence/aot/quantizer/passes/BUCK
Original file line number Diff line number Diff line change
@@ -0,0 +1,16 @@
load("@fbcode_macros//build_defs:build_file_migration.bzl", "fbcode_target", "non_fbcode_target")
load("@fbsource//xplat/executorch/build:runtime_wrapper.bzl", "runtime")

oncall("odai_jarvis")

fbcode_target(_kind = runtime.python_library,
name = "fuse_ops",
srcs = [
"fuse_ops.py",
],
typing = True,
deps = [
"//caffe2:torch",
"//executorch/backends/transforms:quantize_fused_convbn_bias_pass",
],
)
Loading
Loading