You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
# Licensed under the Apache License, Version 2.0 (the "License");
4
+
# you may not use this file except in compliance with the License.
5
+
# You may obtain a copy of the License at
6
+
#
7
+
# https://www.apache.org/licenses/LICENSE-2.0
8
+
#
9
+
# Unless required by applicable law or agreed to in writing, software
10
+
# distributed under the License is distributed on an "AS IS" BASIS,
11
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+
# See the License for the specific language governing permissions and
13
+
# limitations under the License.
14
+
15
+
# This sentinel is a reminder to choose a real run name.
16
+
run_name: ''
17
+
18
+
metrics_file: ""# for testing, local file that stores scalar metrics. If empty, no metrics are written.
19
+
# If true save metrics such as loss and TFLOPS to GCS in {base_output_directory}/{run_name}/metrics/
20
+
write_metrics: True
21
+
22
+
timing_metrics_file: ""# for testing, local file that stores function timing metrics such as state creation, compilation. If empty, no metrics are written.
23
+
write_timing_metrics: True
24
+
25
+
gcs_metrics: False
26
+
# If true save config to GCS in {base_output_directory}/{run_name}/
learning_rate_schedule_steps: -1# By default the length of the schedule is set to the number of steps.
247
+
save_optimizer: False
248
+
249
+
# However you may choose a longer schedule (learning_rate_schedule_steps > steps), in which case the training will end before
250
+
# dropping fully down. Or you may choose a shorter schedule, where the unspecified steps will have a learning rate of 0.
251
+
252
+
# AdamW optimizer parameters
253
+
adam_b1: 0.9# Exponential decay rate to track the first moment of past gradients.
254
+
adam_b2: 0.999# Exponential decay rate to track the second moment of past gradients.
255
+
adam_eps: 1.e-8# A small constant applied to denominator outside of the square root.
256
+
adam_weight_decay: 0# AdamW Weight decay
257
+
max_grad_norm: 1.0
258
+
259
+
enable_profiler: False
260
+
# Skip first n steps for profiling, to omit things like compilation and to give
261
+
# the iteration time a chance to stabilize.
262
+
skip_first_n_steps_for_profiler: 5
263
+
profiler_steps: 10
264
+
265
+
# Enable JAX named scopes for detailed profiling and debugging
266
+
# When enabled, adds named scopes around key operations in transformer and attention layers
267
+
enable_jax_named_scopes: False
268
+
269
+
# Generation parameters
270
+
prompt: "A cat and a dog baking a cake together in a kitchen. The cat is carefully measuring flour, while the dog is stirring the batter with a wooden spoon. The kitchen is cozy, with sunlight streaming through the window."
271
+
prompt_2: "A cat and a dog baking a cake together in a kitchen. The cat is carefully measuring flour, while the dog is stirring the batter with a wooden spoon. The kitchen is cozy, with sunlight streaming through the window."
272
+
negative_prompt: "Bright tones, overexposed, static, blurred details, subtitles, style, works, paintings, images, static, overall gray, worst quality, low quality, JPEG compression residue, ugly, incomplete, extra fingers, poorly drawn hands, poorly drawn faces, deformed, disfigured, misshapen limbs, fused fingers, still picture, messy background, three legs, many people in the background, walking backwards"
273
+
do_classifier_free_guidance: True
274
+
height: 720
275
+
width: 1280
276
+
num_frames: 81
277
+
guidance_scale: 5.0
278
+
flow_shift: 3.0
279
+
280
+
# Based on 3.4. in https://arxiv.org/pdf/2305.08891.pdf
281
+
guidance_rescale: 0.0
282
+
num_inference_steps: 30
283
+
fps: 16
284
+
save_final_checkpoint: False
285
+
286
+
# SDXL Lightning parameters
287
+
lightning_from_pt: True
288
+
# Empty or "ByteDance/SDXL-Lightning" to enable lightning.
289
+
lightning_repo: ""
290
+
# Empty or "sdxl_lightning_4step_unet.safetensors" to enable lightning.
291
+
lightning_ckpt: ""
292
+
293
+
# LoRA parameters
294
+
lora_rank: 64
295
+
# Values are lists to support multiple LoRA loading during inference in the future.
# Shard the range finding operation for quantization. By default this is set to number of slices.
321
+
quantization_local_shard_count: -1
322
+
compile_topology_num_slices: -1# Number of target slices, set to a positive integer.
323
+
use_qwix_quantization: False # Whether to use qwix for quantization. If set to True, the transformer of WAN will be quantized using qwix.
324
+
# Quantization calibration method used for weights and activations. Supported methods can be found in https://github.com/google/qwix/blob/dc2a0770351c740e5ab3cce7c0efe9f7beacce9e/qwix/qconfig.py#L70-L80
325
+
quantization_calibration_method: "absmax"
326
+
qwix_module_path: ".*"
327
+
328
+
# Eval model on per eval_every steps. -1 means don't eval.
329
+
eval_every: -1
330
+
eval_data_dir: ""
331
+
enable_generate_video_for_eval: False # This will increase the used TPU memory.
332
+
eval_max_number_of_samples_in_bucket: 60# The number of samples per bucket for evaluation. This is calculated by num_eval_samples / len(timesteps_list).
0 commit comments