From 07f5d88ac9331d55d27a4e221b93db560e1b1763 Mon Sep 17 00:00:00 2001 From: mi804 <1576993271@qq.com> Date: Mon, 2 Mar 2026 17:37:17 +0800 Subject: [PATCH] update modelid --- .../qwen_image/model_inference/Qwen-Image-Layered-Control-V2.py | 2 +- .../model_inference_low_vram/Qwen-Image-Layered-Control-V2.py | 2 +- .../model_training/lora/Qwen-Image-Layered-Control-V2.sh | 2 +- .../validate_lora/Qwen-Image-Layered-Control-V2.py | 2 +- 4 files changed, 4 insertions(+), 4 deletions(-) diff --git a/examples/qwen_image/model_inference/Qwen-Image-Layered-Control-V2.py b/examples/qwen_image/model_inference/Qwen-Image-Layered-Control-V2.py index 273ee19..ce9968c 100644 --- a/examples/qwen_image/model_inference/Qwen-Image-Layered-Control-V2.py +++ b/examples/qwen_image/model_inference/Qwen-Image-Layered-Control-V2.py @@ -7,7 +7,7 @@ pipe = QwenImagePipeline.from_pretrained( torch_dtype=torch.bfloat16, device="cuda", model_configs=[ - ModelConfig(model_id="Qwen/Qwen-Image-Layered", origin_file_pattern="transformer/diffusion_pytorch_model*.safetensors"), + ModelConfig(model_id="DiffSynth-Studio/Qwen-Image-Layered-Control", origin_file_pattern="transformer/diffusion_pytorch_model*.safetensors"), ModelConfig(model_id="Qwen/Qwen-Image", origin_file_pattern="text_encoder/model*.safetensors"), ModelConfig(model_id="Qwen/Qwen-Image-Layered", origin_file_pattern="vae/diffusion_pytorch_model.safetensors"), ], diff --git a/examples/qwen_image/model_inference_low_vram/Qwen-Image-Layered-Control-V2.py b/examples/qwen_image/model_inference_low_vram/Qwen-Image-Layered-Control-V2.py index 108dab5..70b0fcd 100644 --- a/examples/qwen_image/model_inference_low_vram/Qwen-Image-Layered-Control-V2.py +++ b/examples/qwen_image/model_inference_low_vram/Qwen-Image-Layered-Control-V2.py @@ -18,7 +18,7 @@ pipe = QwenImagePipeline.from_pretrained( torch_dtype=torch.bfloat16, device="cuda", model_configs=[ - ModelConfig(model_id="Qwen/Qwen-Image-Layered", origin_file_pattern="transformer/diffusion_pytorch_model*.safetensors", **vram_config), + ModelConfig(model_id="DiffSynth-Studio/Qwen-Image-Layered-Control", origin_file_pattern="transformer/diffusion_pytorch_model*.safetensors", **vram_config), ModelConfig(model_id="Qwen/Qwen-Image", origin_file_pattern="text_encoder/model*.safetensors", **vram_config), ModelConfig(model_id="Qwen/Qwen-Image-Layered", origin_file_pattern="vae/diffusion_pytorch_model.safetensors", **vram_config), ], diff --git a/examples/qwen_image/model_training/lora/Qwen-Image-Layered-Control-V2.sh b/examples/qwen_image/model_training/lora/Qwen-Image-Layered-Control-V2.sh index 3574dfe..3cce14b 100644 --- a/examples/qwen_image/model_training/lora/Qwen-Image-Layered-Control-V2.sh +++ b/examples/qwen_image/model_training/lora/Qwen-Image-Layered-Control-V2.sh @@ -6,7 +6,7 @@ accelerate launch examples/qwen_image/model_training/train.py \ --data_file_keys "image,layer_input_image,context_image" \ --max_pixels 1048576 \ --dataset_repeat 50 \ - --model_id_with_origin_paths "Qwen/Qwen-Image-Layered:transformer/diffusion_pytorch_model*.safetensors,Qwen/Qwen-Image:text_encoder/model*.safetensors,Qwen/Qwen-Image-Layered:vae/diffusion_pytorch_model.safetensors" \ + --model_id_with_origin_paths "DiffSynth-Studio/Qwen-Image-Layered-Control:transformer/diffusion_pytorch_model*.safetensors,Qwen/Qwen-Image:text_encoder/model*.safetensors,Qwen/Qwen-Image-Layered:vae/diffusion_pytorch_model.safetensors" \ --learning_rate 1e-4 \ --num_epochs 5 \ --remove_prefix_in_ckpt "pipe.dit." \ diff --git a/examples/qwen_image/model_training/validate_lora/Qwen-Image-Layered-Control-V2.py b/examples/qwen_image/model_training/validate_lora/Qwen-Image-Layered-Control-V2.py index 971d35c..3aa7ac4 100644 --- a/examples/qwen_image/model_training/validate_lora/Qwen-Image-Layered-Control-V2.py +++ b/examples/qwen_image/model_training/validate_lora/Qwen-Image-Layered-Control-V2.py @@ -7,7 +7,7 @@ pipe = QwenImagePipeline.from_pretrained( torch_dtype=torch.bfloat16, device="cuda", model_configs=[ - ModelConfig(model_id="Qwen/Qwen-Image-Layered", origin_file_pattern="transformer/diffusion_pytorch_model*.safetensors"), + ModelConfig(model_id="DiffSynth-Studio/Qwen-Image-Layered-Control", origin_file_pattern="transformer/diffusion_pytorch_model*.safetensors"), ModelConfig(model_id="Qwen/Qwen-Image", origin_file_pattern="text_encoder/model*.safetensors"), ModelConfig(model_id="Qwen/Qwen-Image-Layered", origin_file_pattern="vae/diffusion_pytorch_model.safetensors"), ],