You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: src/maxdiffusion/configs/base_wan_14b.yml
+4-2Lines changed: 4 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -341,8 +341,10 @@ quantization: ''
341
341
quantization_local_shard_count: -1
342
342
compile_topology_num_slices: -1# Number of target slices, set to a positive integer.
343
343
use_qwix_quantization: False # Whether to use qwix for quantization. If set to True, the transformer of WAN will be quantized using qwix.
344
-
# Quantization calibration method used for weights and activations. Supported methods can be found in https://github.com/google/qwix/blob/dc2a0770351c740e5ab3cce7c0efe9f7beacce9e/qwix/qconfig.py#L70-L80
345
-
quantization_calibration_method: "absmax"
344
+
# Quantization calibration method used for weights, activations and bwd. Supported methods can be found in https://github.com/google/qwix/blob/dc2a0770351c740e5ab3cce7c0efe9f7beacce9e/qwix/qconfig.py#L70-L80
345
+
weight_quantization_calibration_method: "absmax"
346
+
act_quantization_calibration_method: "absmax"
347
+
bwd_quantization_calibration_method: "absmax"
346
348
qwix_module_path: ".*"
347
349
348
350
# Eval model on per eval_every steps. -1 means don't eval.
0 commit comments