You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: src/maxdiffusion/configs/base_wan_14b.yml
+4-2Lines changed: 4 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -319,8 +319,10 @@ quantization: ''
319
319
quantization_local_shard_count: -1
320
320
compile_topology_num_slices: -1# Number of target slices, set to a positive integer.
321
321
use_qwix_quantization: False # Whether to use qwix for quantization. If set to True, the transformer of WAN will be quantized using qwix.
322
-
# Quantization calibration method used for weights and activations. Supported methods can be found in https://github.com/google/qwix/blob/dc2a0770351c740e5ab3cce7c0efe9f7beacce9e/qwix/qconfig.py#L70-L80
323
-
quantization_calibration_method: "absmax"
322
+
# Quantization calibration method used for weights, activations and bwd. Supported methods can be found in https://github.com/google/qwix/blob/dc2a0770351c740e5ab3cce7c0efe9f7beacce9e/qwix/qconfig.py#L70-L80
323
+
weight_quantization_calibration_method: "absmax"
324
+
act_quantization_calibration_method: "absmax"
325
+
bwd_quantization_calibration_method: "absmax"
324
326
qwix_module_path: ".*"
325
327
326
328
# Eval model on per eval_every steps. -1 means don't eval.
0 commit comments