@@ -445,16 +445,16 @@ def forward(self, x):
445445 Model6GruLinear (), (torch .randn (2 , 16 , 8 ),), 2
446446 ),
447447 "model_7_dwconv_batchnorm_linear" : TransposeCountCase (
448- Model7DwConvBatchNormLinear (), (torch .randn (2 , 8 , 64 ),), 3
448+ Model7DwConvBatchNormLinear (), (torch .randn (2 , 8 , 64 ),), 1
449449 ),
450450 "model_8_conv_batchnorm_maxpool_residual" : TransposeCountCase (
451- Model8ConvBatchNormMaxPoolResidual (), (torch .randn (1 , 8 , 16 , 16 ),), 6
451+ Model8ConvBatchNormMaxPoolResidual (), (torch .randn (1 , 8 , 16 , 16 ),), 4
452452 ),
453453 "model_9_dilated_conv_batchnorm_avgpool_residual" : TransposeCountCase (
454- Model9DilatedConvBatchNormAvgPoolResidual (), (torch .randn (1 , 8 , 16 , 16 ),), 6
454+ Model9DilatedConvBatchNormAvgPoolResidual (), (torch .randn (1 , 8 , 16 , 16 ),), 4
455455 ),
456456 "model_10_dwconv_batchnorm_linear_cat" : TransposeCountCase (
457- Model10DwConvBatchNormLinearCat (), (torch .randn (2 , 8 , 64 ),), 3
457+ Model10DwConvBatchNormLinearCat (), (torch .randn (2 , 8 , 64 ),), 1
458458 ),
459459}
460460
0 commit comments