Skip to content

Commit c44f0e5

Browse files
authored
Logging Fix (#222)
* ltx instruction update * updated whatsnew * updated table of contents * changed order * Fix logging error * renamed json * renamed files
1 parent ee53ee3 commit c44f0e5

3 files changed

Lines changed: 6 additions & 8 deletions

File tree

README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -177,11 +177,11 @@ To generate images, run the following command:
177177
## LTX-Video
178178
- In the folder src/maxdiffusion/models/ltx_video/utils, run:
179179
```bash
180-
python convert_torch_weights_to_jax.py --ckpt_path [LOCAL DIRECTORY FOR WEIGHTS] --transformer_config_path ../xora_v1.2-13B-balanced-128.json
180+
python convert_torch_weights_to_jax.py --ckpt_path [LOCAL DIRECTORY FOR WEIGHTS] --transformer_config_path ../ltxv-13B.json
181181
```
182182
- In the repo folder, run:
183183
```bash
184-
python src/maxdiffusion/generate_ltx_video.py src/maxdiffusion/configs/ltx_video.yml output_dir="[SAME DIRECTORY]" config_path="src/maxdiffusion/models/ltx_video/xora_v1.2-13B-balanced-128.json"
184+
python src/maxdiffusion/generate_ltx_video.py src/maxdiffusion/configs/ltx_video.yml output_dir="[SAME DIRECTORY]" config_path="src/maxdiffusion/models/ltx_video/ltxv-13B.json"
185185
```
186186
- Other generation parameters can be set in ltx_video.yml file.
187187
## Flux

src/maxdiffusion/models/ltx_video/xora_v1.2-13B-balanced-128.json renamed to src/maxdiffusion/models/ltx_video/ltxv-13B.json

File renamed without changes.

src/maxdiffusion/pipelines/ltx_video/ltx_video_pipeline.py

Lines changed: 4 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -60,12 +60,10 @@
6060

6161
def validate_transformer_inputs(prompt_embeds, fractional_coords, latents, encoder_attention_segment_ids):
6262
# Note: reference shape annotated for first pass default inference parameters
63-
max_logging.log("prompts_embeds.shape: ", prompt_embeds.shape, prompt_embeds.dtype) # (3, 256, 4096) float32
64-
max_logging.log("fractional_coords.shape: ", fractional_coords.shape, fractional_coords.dtype) # (3, 3, 3072) float32
65-
max_logging.log("latents.shape: ", latents.shape, latents.dtype) # (1, 3072, 128) float 32
66-
max_logging.log(
67-
"encoder_attention_segment_ids.shape: ", encoder_attention_segment_ids.shape, encoder_attention_segment_ids.dtype
68-
) # (3, 256) int32
63+
max_logging.log(f"prompts_embeds.shape: {prompt_embeds.shape}") # (3, 256, 4096) float32
64+
max_logging.log(f"fractional_coords.shape: {fractional_coords.shape}") # (3, 3, 3072) float32
65+
max_logging.log(f"latents.shape: {latents.shape}") # (1, 3072, 128) float 32
66+
max_logging.log(f"encoder_attention_segment_ids.shape: {encoder_attention_segment_ids.shape}") # (3, 256) int32
6967

7068

7169
class LTXVideoPipeline:

0 commit comments

Comments
 (0)