Skip to content

Commit 9042f36

Browse files
authored
Merge branch 'main' into cuda-graph
2 parents 10e7aad + 2eaa16c commit 9042f36

3 files changed

Lines changed: 97 additions & 180 deletions

File tree

backends/nxp/runtime/NeutronBackend.cpp

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -433,12 +433,12 @@ class NeutronBackend final : public PyTorchBackendInterface {
433433

434434
if (is_channels_last_dim_order(dim_order, arg.dim())) {
435435
// The tensor is already permuted.
436-
ET_LOG(Info, "Using channels last dim order for input %d.\n", i);
436+
ET_LOG(Debug, "Using channels last dim order for input %d.\n", i);
437437
cfg->dcfg.inputs[i] = arg.const_data_ptr();
438438
} else if (is_contiguous_dim_order(dim_order, arg.dim())) {
439439
// Transpose the data to channels last.
440440

441-
ET_LOG(Info, "Transposing input %d to channels last.\n", i);
441+
ET_LOG(Debug, "Transposing input %d to channels last.\n", i);
442442

443443
// Allocate buffer, the allocator is reset after each PTE instruction.
444444
void* buffer = context.allocate(arg.nbytes(), 16);
@@ -542,10 +542,10 @@ class NeutronBackend final : public PyTorchBackendInterface {
542542
if (is_channels_last_dim_order(dim_order, arg.dim())) {
543543
// The rest of the model expects the `channels_last` dim order, which
544544
// the data already matches.
545-
ET_LOG(Info, "Using channels last dim order for output %d.\n", i);
545+
ET_LOG(Debug, "Using channels last dim order for output %d.\n", i);
546546
} else if (is_contiguous_dim_order(dim_order, arg.dim())) {
547547
// Transpose the data to channels first.
548-
ET_LOG(Info, "Transposing output %d to channels first.\n", i);
548+
ET_LOG(Debug, "Transposing output %d to channels first.\n", i);
549549
transposeOutput(
550550
cfg->dcfg.outputs[i],
551551
arg.mutable_data_ptr(),

0 commit comments

Comments
 (0)