Skip to content

Commit c73595a

Browse files
Merge pull request #3645 from AI-Hypercomputer:chengnuojin-correct-logical
PiperOrigin-RevId: 898309365
2 parents f216197 + 07f408e commit c73595a

2 files changed

Lines changed: 3 additions & 3 deletions

File tree

src/maxtext/configs/base.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -479,7 +479,7 @@ logical_axis_rules: [
479479
['activation_mlp_moe', ['tensor', 'tensor_transpose', 'tensor_sequence']],
480480
['activation_kv', ['tensor', 'tensor_transpose', 'tensor_sequence']],
481481
['activation_prefill_kv_batch', ['data', 'fsdp', 'fsdp_transpose', 'expert']],
482-
['activation_kv_batch', ['data', 'fsdp', 'fsdp_transpose']],
482+
['activation_kv_batch', ['data', 'fsdp', 'fsdp_transpose', 'expert']],
483483
['activation_kv_head_dim', ['tensor', 'tensor_transpose', 'tensor_sequence']],
484484
['activation_vocab', ['tensor', 'tensor_transpose', 'tensor_sequence']],
485485
['activation_vocab', ['tensor', 'tensor_transpose']],
@@ -969,7 +969,7 @@ xprof_e2e_enable_fw_power_level_event: False
969969
xprof_e2e_enable_fw_thermal_event: False
970970
profile_power_events: False # Set to True to enable TPU-specific power/thermal profiling events. Defaults to False to avoid breaking GPU xplane tracing.
971971

972-
log_config: False # Prints the config (after defaults have been set by pyconfig logic)
972+
log_config: True # Prints the config (after defaults have been set by pyconfig logic)
973973
debug_sharding: False # Prints model weights sharding info
974974

975975
# Checkpoint Structured logging

src/maxtext/configs/custom_mesh_and_rule/pipeline-large-moe.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,7 @@ logical_axis_rules: [
4646
['activation_mlp', ['tensor']],
4747
['activation_mlp_moe', ['tensor']],
4848
['activation_kv', ['tensor']],
49-
['activation_kv_batch', ['data', 'fsdp']],
49+
['activation_kv_batch', ['data', 'fsdp', 'expert']],
5050
['activation_kv_head_dim', ['tensor']],
5151
['activation_vocab', ['tensor']],
5252
['activation_stage', 'stage'],

0 commit comments

Comments
 (0)