mirror of
https://github.com/modelscope/DiffSynth-Studio.git
synced 2026-03-24 01:48:13 +00:00
update qwen-image-edit training script
This commit is contained in:
@@ -1,6 +1,6 @@
|
|||||||
accelerate launch --config_file examples/qwen_image/model_training/full/accelerate_config_zero2offload.yaml examples/qwen_image/model_training/train.py \
|
accelerate launch --config_file examples/qwen_image/model_training/full/accelerate_config_zero2offload.yaml examples/qwen_image/model_training/train.py \
|
||||||
--dataset_base_path data/example_image_dataset \
|
--dataset_base_path data/example_image_dataset \
|
||||||
--dataset_metadata_path data/example_image_dataset/metadata_edit.csv \
|
--dataset_metadata_path data/example_image_dataset/metadata_qwen_imgae_edit_multi.json \
|
||||||
--data_file_keys "image,edit_image" \
|
--data_file_keys "image,edit_image" \
|
||||||
--extra_inputs "edit_image" \
|
--extra_inputs "edit_image" \
|
||||||
--max_pixels 1048576 \
|
--max_pixels 1048576 \
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
accelerate launch examples/qwen_image/model_training/train.py \
|
accelerate launch examples/qwen_image/model_training/train.py \
|
||||||
--dataset_base_path data/example_image_dataset \
|
--dataset_base_path data/example_image_dataset \
|
||||||
--dataset_metadata_path data/example_image_dataset/metadata_edit.csv \
|
--dataset_metadata_path data/example_image_dataset/metadata_qwen_imgae_edit_multi.json \
|
||||||
--data_file_keys "image,edit_image" \
|
--data_file_keys "image,edit_image" \
|
||||||
--extra_inputs "edit_image" \
|
--extra_inputs "edit_image" \
|
||||||
--max_pixels 1048576 \
|
--max_pixels 1048576 \
|
||||||
|
|||||||
@@ -17,7 +17,10 @@ pipe = QwenImagePipeline.from_pretrained(
|
|||||||
state_dict = load_state_dict("models/train/Qwen-Image-Edit-2509_full/epoch-1.safetensors")
|
state_dict = load_state_dict("models/train/Qwen-Image-Edit-2509_full/epoch-1.safetensors")
|
||||||
pipe.dit.load_state_dict(state_dict)
|
pipe.dit.load_state_dict(state_dict)
|
||||||
|
|
||||||
prompt = "将裙子改为粉色"
|
prompt = "Change the color of the dress in Figure 1 to the color shown in Figure 2."
|
||||||
image = Image.open("data/example_image_dataset/edit/image1.jpg").resize((1024, 1024))
|
images = [
|
||||||
image = pipe(prompt, edit_image=image, seed=0, num_inference_steps=40, height=1024, width=1024)
|
Image.open("data/example_image_dataset/edit/image1.jpg").resize((1024, 1024)),
|
||||||
image.save(f"image.jpg")
|
Image.open("data/example_image_dataset/edit/image_color.jpg").resize((1024, 1024)),
|
||||||
|
]
|
||||||
|
image = pipe(prompt, edit_image=images, seed=123, num_inference_steps=40, height=1024, width=1024)
|
||||||
|
image.save("image.jpg")
|
||||||
|
|||||||
@@ -15,7 +15,10 @@ pipe = QwenImagePipeline.from_pretrained(
|
|||||||
)
|
)
|
||||||
pipe.load_lora(pipe.dit, "models/train/Qwen-Image-Edit-2509_lora/epoch-4.safetensors")
|
pipe.load_lora(pipe.dit, "models/train/Qwen-Image-Edit-2509_lora/epoch-4.safetensors")
|
||||||
|
|
||||||
prompt = "将裙子改为粉色"
|
prompt = "Change the color of the dress in Figure 1 to the color shown in Figure 2."
|
||||||
image = Image.open("data/example_image_dataset/edit/image1.jpg").resize((1024, 1024))
|
images = [
|
||||||
image = pipe(prompt, edit_image=image, seed=0, num_inference_steps=40, height=1024, width=1024)
|
Image.open("data/example_image_dataset/edit/image1.jpg").resize((1024, 1024)),
|
||||||
image.save(f"image.jpg")
|
Image.open("data/example_image_dataset/edit/image_color.jpg").resize((1024, 1024)),
|
||||||
|
]
|
||||||
|
image = pipe(prompt, edit_image=images, seed=123, num_inference_steps=40, height=1024, width=1024)
|
||||||
|
image.save("image.jpg")
|
||||||
|
|||||||
Reference in New Issue
Block a user