@@ -433,12 +433,12 @@ class NeutronBackend final : public PyTorchBackendInterface {
433433
434434 if (is_channels_last_dim_order (dim_order, arg.dim ())) {
435435 // The tensor is already permuted.
436- ET_LOG (Info , " Using channels last dim order for input %d.\n " , i);
436+ ET_LOG (Debug , " Using channels last dim order for input %d.\n " , i);
437437 cfg->dcfg .inputs [i] = arg.const_data_ptr ();
438438 } else if (is_contiguous_dim_order (dim_order, arg.dim ())) {
439439 // Transpose the data to channels last.
440440
441- ET_LOG (Info , " Transposing input %d to channels last.\n " , i);
441+ ET_LOG (Debug , " Transposing input %d to channels last.\n " , i);
442442
443443 // Allocate buffer, the allocator is reset after each PTE instruction.
444444 void * buffer = context.allocate (arg.nbytes (), 16 );
@@ -542,10 +542,10 @@ class NeutronBackend final : public PyTorchBackendInterface {
542542 if (is_channels_last_dim_order (dim_order, arg.dim ())) {
543543 // The rest of the model expects the `channels_last` dim order, which
544544 // the data already matches.
545- ET_LOG (Info , " Using channels last dim order for output %d.\n " , i);
545+ ET_LOG (Debug , " Using channels last dim order for output %d.\n " , i);
546546 } else if (is_contiguous_dim_order (dim_order, arg.dim ())) {
547547 // Transpose the data to channels first.
548- ET_LOG (Info , " Transposing output %d to channels first.\n " , i);
548+ ET_LOG (Debug , " Transposing output %d to channels first.\n " , i);
549549 transposeOutput (
550550 cfg->dcfg .outputs [i],
551551 arg.mutable_data_ptr (),
0 commit comments