mirror of
https://github.com/modelscope/DiffSynth-Studio.git
synced 2026-03-18 22:08:13 +00:00
40 lines
1.9 KiB
Bash
40 lines
1.9 KiB
Bash
accelerate launch examples/z_image/model_training/train.py \
|
|
--dataset_base_path data/example_image_dataset \
|
|
--dataset_metadata_path data/example_image_dataset/metadata.csv \
|
|
--max_pixels 1048576 \
|
|
--dataset_repeat 50 \
|
|
--model_id_with_origin_paths "Tongyi-MAI/Z-Image-Turbo:transformer/*.safetensors,Tongyi-MAI/Z-Image-Turbo:text_encoder/*.safetensors,Tongyi-MAI/Z-Image-Turbo:vae/diffusion_pytorch_model.safetensors" \
|
|
--learning_rate 1e-4 \
|
|
--num_epochs 5 \
|
|
--remove_prefix_in_ckpt "pipe.dit." \
|
|
--output_path "./models/train/Z-Image-Turbo_lora" \
|
|
--lora_base_model "dit" \
|
|
--lora_target_modules "to_q,to_k,to_v,to_out.0,w1,w2,w3" \
|
|
--lora_rank 32 \
|
|
--use_gradient_checkpointing \
|
|
--dataset_num_workers 8
|
|
|
|
|
|
# Z-Image-Turbo is a distilled model.
|
|
# After training, it loses its distillation-based acceleration capability,
|
|
# leading to degraded generation quality at fewer inference steps.
|
|
# This issue can be mitigated by using a pre-trained LoRA model to assist the training process.
|
|
|
|
# accelerate launch examples/z_image/model_training/train.py \
|
|
# --dataset_base_path data/example_image_dataset \
|
|
# --dataset_metadata_path data/example_image_dataset/metadata.csv \
|
|
# --max_pixels 1048576 \
|
|
# --dataset_repeat 50 \
|
|
# --model_id_with_origin_paths "Tongyi-MAI/Z-Image-Turbo:transformer/*.safetensors,Tongyi-MAI/Z-Image-Turbo:text_encoder/*.safetensors,Tongyi-MAI/Z-Image-Turbo:vae/diffusion_pytorch_model.safetensors" \
|
|
# --learning_rate 1e-4 \
|
|
# --num_epochs 5 \
|
|
# --remove_prefix_in_ckpt "pipe.dit." \
|
|
# --output_path "./models/train/Z-Image-Turbo_lora" \
|
|
# --lora_base_model "dit" \
|
|
# --lora_target_modules "to_q,to_k,to_v,to_out.0,w1,w2,w3" \
|
|
# --lora_rank 32 \
|
|
# --preset_lora_path "models/ostris/zimage_turbo_training_adapter/zimage_turbo_training_adapter_v1.safetensors" \
|
|
# --preset_lora_model "dit" \
|
|
# --use_gradient_checkpointing \
|
|
# --dataset_num_workers 8
|