Files
DiffSynth-Studio/examples/wanvideo/README.md
2025-06-06 14:58:41 +08:00

33 lines
1.7 KiB
Markdown

* dataset
* `--dataset_base_path`: Base path of the Dataset.
* `--dataset_metadata_path`: Metadata path of the Dataset.
* `--height`: Image or video height. Leave `height` and `width` None to enable dynamic resolution.
* `--width`: Image or video width. Leave `height` and `width` None to enable dynamic resolution.
* `--num_frames`: Number of frames in each video. The frames are sampled from the prefix.
* `--data_file_keys`: Data file keys in metadata. Separated by commas.
* `--dataset_repeat`: Number of times the dataset is repeated in each epoch.
* Model
* `--model_paths`: Model paths to be loaded. JSON format.
* `--model_id_with_origin_paths`: Model ID with original path, e.g., Wan-AI/Wan2.1-T2V-1.3B:diffusion_pytorch_model*.safetensors. Separated by commas.
* Training
* `--learning_rate`: Learning rate.
* `--num_epochs`: Number of epochs.
* `--output_path`: Save path.
* `--remove_prefix_in_ckpt`: Remove prefix in ckpt.
* Trainable module
* `--trainable_models`: Trainable models, e.g., dit, vae, text_encoder.
* `--lora_base_model`: Add LoRA on which model.
* `--lora_target_modules`: Add LoRA on which layer.
* `--lora_rank`: LoRA rank.
* Extra model input
* `--input_contains_input_image`: Model input contains `input_image`
* `--input_contains_end_image`: Model input contains `end_image`.
* `--input_contains_control_video`: Model input contains `control_video`.
* `--input_contains_reference_image`: Model input contains `reference_image`.
* `--input_contains_vace_video`: Model input contains `vace_video`.
* `--input_contains_vace_reference_image`: Model input contains `vace_reference_image`.
* `--input_contains_motion_bucket_id`: Model input contains `motion_bucket_id`.