66
77from typing import Tuple
88
9- import pytest
10-
119import torch
1210from executorch .backends .arm ._passes import (
1311 ConvertInt64ConstOpsToInt32Pass ,
@@ -41,20 +39,17 @@ class TestCLIPTextModelWithProjection:
4139 ops_after_partitioner_FP = {
4240 "executorch_exir_dialects_edge__ops_aten_argmax_default" : 1 ,
4341 "executorch_exir_dialects_edge__ops_dim_order_ops__to_dim_order_copy_default" : 2 ,
44- "torch.ops.higher_order.executorch_call_delegate" : 1 ,
42+ "torch.ops.higher_order.executorch_call_delegate" : 2 ,
4543 }
4644
4745 ops_after_partitioner_INT = {
4846 "executorch_exir_dialects_edge__ops_aten_argmax_default" : 1 ,
4947 "executorch_exir_dialects_edge__ops_dim_order_ops__to_dim_order_copy_default" : 2 ,
48+ "torch.ops.higher_order.executorch_call_delegate" : 1 ,
5049 }
5150
52- ops_after_partitioner_vgf_quantize = ops_after_partitioner_FP
53- ops_after_partitioner_vgf_no_quantize = {
54- "executorch_exir_dialects_edge__ops_aten_argmax_default" : 1 ,
55- "executorch_exir_dialects_edge__ops_dim_order_ops__to_dim_order_copy_default" : 2 ,
56- "torch.ops.higher_order.executorch_call_delegate" : 2 ,
57- }
51+ ops_after_partitioner_vgf_quantize = ops_after_partitioner_INT
52+ ops_after_partitioner_vgf_no_quantize = ops_after_partitioner_FP
5853
5954 def _prepare_inputs (
6055 self ,
@@ -80,9 +75,6 @@ def prepare_model_and_inputs(self):
8075 return text_encoder_model , text_encoder_model_inputs
8176
8277
83- @pytest .mark .xfail (
84- reason = "MLETORCH-1601: Delegate output order mismatch from TOSA reference model."
85- )
8678def test_clip_text_with_projection_tosa_FP ():
8779 text_encoder_model , text_encoder_model_inputs = (
8880 TestCLIPTextModelWithProjection ().prepare_model_and_inputs ()
0 commit comments