From b2554db100255f75eb4a4cbb7c9780a90a21a775 Mon Sep 17 00:00:00 2001 From: mi804 <1576993271@qq.com> Date: Fri, 1 Aug 2025 17:13:45 +0800 Subject: [PATCH] fix krea typo --- README.md | 4 ++-- README_zh.md | 4 ++-- examples/flux/README.md | 2 +- examples/flux/README_zh.md | 2 +- .../{FLUX.1-Kera-dev.py => FLUX.1-Krea-dev.py} | 4 ++-- .../{FLUX.1-Kera-dev.py => FLUX.1-Krea-dev.py} | 4 ++-- .../full/{FLUX.1-Kera-dev.sh => FLUX.1-Krea-dev.sh} | 2 +- .../lora/{FLUX.1-Kera-dev.sh => FLUX.1-Krea-dev.sh} | 2 +- .../validate_full/{FLUX.1-Kera-dev.py => FLUX.1-Krea-dev.py} | 4 ++-- .../validate_lora/{FLUX.1-Kera-dev.py => FLUX.1-Krea-dev.py} | 4 ++-- 10 files changed, 16 insertions(+), 16 deletions(-) rename examples/flux/model_inference/{FLUX.1-Kera-dev.py => FLUX.1-Krea-dev.py} (94%) rename examples/flux/model_inference_low_vram/{FLUX.1-Kera-dev.py => FLUX.1-Krea-dev.py} (95%) rename examples/flux/model_training/full/{FLUX.1-Kera-dev.sh => FLUX.1-Krea-dev.sh} (92%) rename examples/flux/model_training/lora/{FLUX.1-Kera-dev.sh => FLUX.1-Krea-dev.sh} (93%) rename examples/flux/model_training/validate_full/{FLUX.1-Kera-dev.py => FLUX.1-Krea-dev.py} (87%) rename examples/flux/model_training/validate_lora/{FLUX.1-Kera-dev.py => FLUX.1-Krea-dev.py} (86%) diff --git a/README.md b/README.md index 6b4f0c0..1ec6947 100644 --- a/README.md +++ b/README.md @@ -90,7 +90,7 @@ image.save("image.jpg") | Model ID | Extra Parameters | Inference | Low VRAM Inference | Full Training | Validate After Full Training | LoRA Training | Validate After LoRA Training | |-|-|-|-|-|-|-|-| |[FLUX.1-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-dev)||[code](./examples/flux/model_inference/FLUX.1-dev.py)|[code](./examples/flux/model_inference_low_vram/FLUX.1-dev.py)|[code](./examples/flux/model_training/full/FLUX.1-dev.sh)|[code](./examples/flux/model_training/validate_full/FLUX.1-dev.py)|[code](./examples/flux/model_training/lora/FLUX.1-dev.sh)|[code](./examples/flux/model_training/validate_lora/FLUX.1-dev.py)| -|[FLUX.1-Kera-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-Krea-dev)||[code](./examples/flux/model_inference/FLUX.1-Kera-dev.py)|[code](./examples/flux/model_inference_low_vram/FLUX.1-Kera-dev.py)|[code](./examples/flux/model_training/full/FLUX.1-Kera-dev.sh)|[code](./examples/flux/model_training/validate_full/FLUX.1-Kera-dev.py)|[code](./examples/flux/model_training/lora/FLUX.1-Kera-dev.sh)|[code](./examples/flux/model_training/validate_lora/FLUX.1-Kera-dev.py)| +|[FLUX.1-Krea-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-Krea-dev)||[code](./examples/flux/model_inference/FLUX.1-Krea-dev.py)|[code](./examples/flux/model_inference_low_vram/FLUX.1-Krea-dev.py)|[code](./examples/flux/model_training/full/FLUX.1-Krea-dev.sh)|[code](./examples/flux/model_training/validate_full/FLUX.1-Krea-dev.py)|[code](./examples/flux/model_training/lora/FLUX.1-Krea-dev.sh)|[code](./examples/flux/model_training/validate_lora/FLUX.1-Krea-dev.py)| |[FLUX.1-Kontext-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-Kontext-dev)|`kontext_images`|[code](./examples/flux/model_inference/FLUX.1-Kontext-dev.py)|[code](./examples/flux/model_inference_low_vram/FLUX.1-Kontext-dev.py)|[code](./examples/flux/model_training/full/FLUX.1-Kontext-dev.sh)|[code](./examples/flux/model_training/validate_full/FLUX.1-Kontext-dev.py)|[code](./examples/flux/model_training/lora/FLUX.1-Kontext-dev.sh)|[code](./examples/flux/model_training/validate_lora/FLUX.1-Kontext-dev.py)| |[FLUX.1-dev-Controlnet-Inpainting-Beta](https://www.modelscope.cn/models/alimama-creative/FLUX.1-dev-Controlnet-Inpainting-Beta)|`controlnet_inputs`|[code](./examples/flux/model_inference/FLUX.1-dev-Controlnet-Inpainting-Beta.py)|[code](./examples/flux/model_inference_low_vram/FLUX.1-dev-Controlnet-Inpainting-Beta.py)|[code](./examples/flux/model_training/full/FLUX.1-dev-Controlnet-Inpainting-Beta.sh)|[code](./examples/flux/model_training/validate_full/FLUX.1-dev-Controlnet-Inpainting-Beta.py)|[code](./examples/flux/model_training/lora/FLUX.1-dev-Controlnet-Inpainting-Beta.sh)|[code](./examples/flux/model_training/validate_lora/FLUX.1-dev-Controlnet-Inpainting-Beta.py)| |[FLUX.1-dev-Controlnet-Union-alpha](https://www.modelscope.cn/models/InstantX/FLUX.1-dev-Controlnet-Union-alpha)|`controlnet_inputs`|[code](./examples/flux/model_inference/FLUX.1-dev-Controlnet-Union-alpha.py)|[code](./examples/flux/model_inference_low_vram/FLUX.1-dev-Controlnet-Union-alpha.py)|[code](./examples/flux/model_training/full/FLUX.1-dev-Controlnet-Union-alpha.sh)|[code](./examples/flux/model_training/validate_full/FLUX.1-dev-Controlnet-Union-alpha.py)|[code](./examples/flux/model_training/lora/FLUX.1-dev-Controlnet-Union-alpha.sh)|[code](./examples/flux/model_training/validate_lora/FLUX.1-dev-Controlnet-Union-alpha.py)| @@ -321,7 +321,7 @@ https://github.com/Artiprocher/DiffSynth-Studio/assets/35051019/59fb2f7b-8de0-44 ## Update History -- **August 1, 2025** [FLUX.1-Kera-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-Krea-dev) with a focus on aesthetic photography is comprehensively supported, including low-GPU-memory layer-by-layer offload, LoRA training and full training. See [./examples/flux/](./examples/flux/). +- **August 1, 2025** [FLUX.1-Krea-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-Krea-dev) with a focus on aesthetic photography is comprehensively supported, including low-GPU-memory layer-by-layer offload, LoRA training and full training. See [./examples/flux/](./examples/flux/). - **July 28, 2025** 🔥🔥🔥 With the open-sourcing of Wan 2.2, we immediately provided comprehensive support, including low-GPU-memory layer-by-layer offload, FP8 quantization, sequence parallelism, LoRA training, full training. See [./examples/wanvideo/](./examples/wanvideo/). diff --git a/README_zh.md b/README_zh.md index e704b0d..5ea73b1 100644 --- a/README_zh.md +++ b/README_zh.md @@ -150,7 +150,7 @@ save_video(video, "video1.mp4", fps=15, quality=5) |模型 ID|额外参数|推理|全量训练|全量训练后验证|LoRA 训练|LoRA 训练后验证| |-|-|-|-|-|-|-| |[Wan-AI/Wan2.2-I2V-A14B](https://modelscope.cn/models/Wan-AI/Wan2.2-I2V-A14B)|`input_image`|[code](./examples/wanvideo/model_inference/Wan2.2-I2V-A14B.py)|[code](./examples/wanvideo/model_training/full/Wan2.2-I2V-A14B.sh)|[code](./examples/wanvideo/model_training/validate_full/Wan2.2-I2V-A14B.py)|[code](./examples/wanvideo/model_training/lora/Wan2.2-I2V-A14B.sh)|[code](./examples/wanvideo/model_training/validate_lora/Wan2.2-I2V-A14B.py)| -|[FLUX.1-Kera-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-Krea-dev)||[code](./examples/flux/model_inference/FLUX.1-Kera-dev.py)|[code](./examples/flux/model_inference_low_vram/FLUX.1-Kera-dev.py)|[code](./examples/flux/model_training/full/FLUX.1-Kera-dev.sh)|[code](./examples/flux/model_training/validate_full/FLUX.1-Kera-dev.py)|[code](./examples/flux/model_training/lora/FLUX.1-Kera-dev.sh)|[code](./examples/flux/model_training/validate_lora/FLUX.1-Kera-dev.py)| +|[FLUX.1-Krea-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-Krea-dev)||[code](./examples/flux/model_inference/FLUX.1-Krea-dev.py)|[code](./examples/flux/model_inference_low_vram/FLUX.1-Krea-dev.py)|[code](./examples/flux/model_training/full/FLUX.1-Krea-dev.sh)|[code](./examples/flux/model_training/validate_full/FLUX.1-Krea-dev.py)|[code](./examples/flux/model_training/lora/FLUX.1-Krea-dev.sh)|[code](./examples/flux/model_training/validate_lora/FLUX.1-Krea-dev.py)| |[Wan-AI/Wan2.2-T2V-A14B](https://modelscope.cn/models/Wan-AI/Wan2.2-T2V-A14B)||[code](./examples/wanvideo/model_inference/Wan2.2-T2V-A14B.py)|[code](./examples/wanvideo/model_training/full/Wan2.2-T2V-A14B.sh)|[code](./examples/wanvideo/model_training/validate_full/Wan2.2-T2V-A14B.py)|[code](./examples/wanvideo/model_training/lora/Wan2.2-T2V-A14B.sh)|[code](./examples/wanvideo/model_training/validate_lora/Wan2.2-T2V-A14B.py)| |[Wan-AI/Wan2.2-TI2V-5B](https://modelscope.cn/models/Wan-AI/Wan2.2-TI2V-5B)|`input_image`|[code](./examples/wanvideo/model_inference/Wan2.2-TI2V-5B.py)|[code](./examples/wanvideo/model_training/full/Wan2.2-TI2V-5B.sh)|[code](./examples/wanvideo/model_training/validate_full/Wan2.2-TI2V-5B.py)|[code](./examples/wanvideo/model_training/lora/Wan2.2-TI2V-5B.sh)|[code](./examples/wanvideo/model_training/validate_lora/Wan2.2-TI2V-5B.py)| |[Wan-AI/Wan2.1-T2V-1.3B](https://modelscope.cn/models/Wan-AI/Wan2.1-T2V-1.3B)||[code](./examples/wanvideo/model_inference/Wan2.1-T2V-1.3B.py)|[code](./examples/wanvideo/model_training/full/Wan2.1-T2V-1.3B.sh)|[code](./examples/wanvideo/model_training/validate_full/Wan2.1-T2V-1.3B.py)|[code](./examples/wanvideo/model_training/lora/Wan2.1-T2V-1.3B.sh)|[code](./examples/wanvideo/model_training/validate_lora/Wan2.1-T2V-1.3B.py)| @@ -337,7 +337,7 @@ https://github.com/Artiprocher/DiffSynth-Studio/assets/35051019/59fb2f7b-8de0-44 ## 更新历史 -- **2025年8月1日** [FLUX.1-Kera-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-Krea-dev) 开源,这是一个专注于美学摄影的文生图模型。我们第一时间提供了全方位支持,包括低显存逐层 offload、LoRA 训练、全量训练。详细信息请参考 [./examples/flux/](./examples/flux/)。 +- **2025年8月1日** [FLUX.1-Krea-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-Krea-dev) 开源,这是一个专注于美学摄影的文生图模型。我们第一时间提供了全方位支持,包括低显存逐层 offload、LoRA 训练、全量训练。详细信息请参考 [./examples/flux/](./examples/flux/)。 - **2025年7月28日** 🔥🔥🔥 Wan 2.2 开源,我们第一时间提供了全方位支持,包括低显存逐层 offload、FP8 量化、序列并行、LoRA 训练、全量训练。详细信息请参考 [./examples/wanvideo/](./examples/wanvideo/)。 diff --git a/examples/flux/README.md b/examples/flux/README.md index 0ddad54..7361907 100644 --- a/examples/flux/README.md +++ b/examples/flux/README.md @@ -44,7 +44,7 @@ image.save("image.jpg") |Model ID|Extra Args|Inference|Low VRAM Inference|Full Training|Validation after Full Training|LoRA Training|Validation after LoRA Training| |-|-|-|-|-|-|-|-| |[FLUX.1-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-dev)||[code](./model_inference/FLUX.1-dev.py)|[code](./model_inference_low_vram/FLUX.1-dev.py)|[code](./model_training/full/FLUX.1-dev.sh)|[code](./model_training/validate_full/FLUX.1-dev.py)|[code](./model_training/lora/FLUX.1-dev.sh)|[code](./model_training/validate_lora/FLUX.1-dev.py)| -|[FLUX.1-Kera-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-Krea-dev)||[code](./model_inference/FLUX.1-Kera-dev.py)|[code](./model_inference_low_vram/FLUX.1-Kera-dev.py)|[code](./model_training/full/FLUX.1-Kera-dev.sh)|[code](./model_training/validate_full/FLUX.1-Kera-dev.py)|[code](./model_training/lora/FLUX.1-Kera-dev.sh)|[code](./model_training/validate_lora/FLUX.1-Kera-dev.py)| +|[FLUX.1-Krea-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-Krea-dev)||[code](./model_inference/FLUX.1-Krea-dev.py)|[code](./model_inference_low_vram/FLUX.1-Krea-dev.py)|[code](./model_training/full/FLUX.1-Krea-dev.sh)|[code](./model_training/validate_full/FLUX.1-Krea-dev.py)|[code](./model_training/lora/FLUX.1-Krea-dev.sh)|[code](./model_training/validate_lora/FLUX.1-Krea-dev.py)| |[FLUX.1-Kontext-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-Kontext-dev)|`kontext_images`|[code](./model_inference/FLUX.1-Kontext-dev.py)|[code](./model_inference_low_vram/FLUX.1-Kontext-dev.py)|[code](./model_training/full/FLUX.1-Kontext-dev.sh)|[code](./model_training/validate_full/FLUX.1-Kontext-dev.py)|[code](./model_training/lora/FLUX.1-Kontext-dev.sh)|[code](./model_training/validate_lora/FLUX.1-Kontext-dev.py)| |[FLUX.1-dev-Controlnet-Inpainting-Beta](https://www.modelscope.cn/models/alimama-creative/FLUX.1-dev-Controlnet-Inpainting-Beta)|`controlnet_inputs`|[code](./model_inference/FLUX.1-dev-Controlnet-Inpainting-Beta.py)|[code](./model_inference_low_vram/FLUX.1-dev-Controlnet-Inpainting-Beta.py)|[code](./model_training/full/FLUX.1-dev-Controlnet-Inpainting-Beta.sh)|[code](./model_training/validate_full/FLUX.1-dev-Controlnet-Inpainting-Beta.py)|[code](./model_training/lora/FLUX.1-dev-Controlnet-Inpainting-Beta.sh)|[code](./model_training/validate_lora/FLUX.1-dev-Controlnet-Inpainting-Beta.py)| |[FLUX.1-dev-Controlnet-Union-alpha](https://www.modelscope.cn/models/InstantX/FLUX.1-dev-Controlnet-Union-alpha)|`controlnet_inputs`|[code](./model_inference/FLUX.1-dev-Controlnet-Union-alpha.py)|[code](./model_inference_low_vram/FLUX.1-dev-Controlnet-Union-alpha.py)|[code](./model_training/full/FLUX.1-dev-Controlnet-Union-alpha.sh)|[code](./model_training/validate_full/FLUX.1-dev-Controlnet-Union-alpha.py)|[code](./model_training/lora/FLUX.1-dev-Controlnet-Union-alpha.sh)|[code](./model_training/validate_lora/FLUX.1-dev-Controlnet-Union-alpha.py)| diff --git a/examples/flux/README_zh.md b/examples/flux/README_zh.md index 02f1e01..6bb3100 100644 --- a/examples/flux/README_zh.md +++ b/examples/flux/README_zh.md @@ -44,7 +44,7 @@ image.save("image.jpg") |模型 ID|额外参数|推理|低显存推理|全量训练|全量训练后验证|LoRA 训练|LoRA 训练后验证| |-|-|-|-|-|-|-|-| |[FLUX.1-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-dev)||[code](./model_inference/FLUX.1-dev.py)|[code](./model_inference_low_vram/FLUX.1-dev.py)|[code](./model_training/full/FLUX.1-dev.sh)|[code](./model_training/validate_full/FLUX.1-dev.py)|[code](./model_training/lora/FLUX.1-dev.sh)|[code](./model_training/validate_lora/FLUX.1-dev.py)| -|[FLUX.1-Kera-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-Krea-dev)||[code](./model_inference/FLUX.1-Kera-dev.py)|[code](./model_inference_low_vram/FLUX.1-Kera-dev.py)|[code](./model_training/full/FLUX.1-Kera-dev.sh)|[code](./model_training/validate_full/FLUX.1-Kera-dev.py)|[code](./model_training/lora/FLUX.1-Kera-dev.sh)|[code](./model_training/validate_lora/FLUX.1-Kera-dev.py)| +|[FLUX.1-Krea-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-Krea-dev)||[code](./model_inference/FLUX.1-Krea-dev.py)|[code](./model_inference_low_vram/FLUX.1-Krea-dev.py)|[code](./model_training/full/FLUX.1-Krea-dev.sh)|[code](./model_training/validate_full/FLUX.1-Krea-dev.py)|[code](./model_training/lora/FLUX.1-Krea-dev.sh)|[code](./model_training/validate_lora/FLUX.1-Krea-dev.py)| |[FLUX.1-Kontext-dev](https://www.modelscope.cn/models/black-forest-labs/FLUX.1-Kontext-dev)|`kontext_images`|[code](./model_inference/FLUX.1-Kontext-dev.py)|[code](./model_inference_low_vram/FLUX.1-Kontext-dev.py)|[code](./model_training/full/FLUX.1-Kontext-dev.sh)|[code](./model_training/validate_full/FLUX.1-Kontext-dev.py)|[code](./model_training/lora/FLUX.1-Kontext-dev.sh)|[code](./model_training/validate_lora/FLUX.1-Kontext-dev.py)| |[FLUX.1-dev-Controlnet-Inpainting-Beta](https://www.modelscope.cn/models/alimama-creative/FLUX.1-dev-Controlnet-Inpainting-Beta)|`controlnet_inputs`|[code](./model_inference/FLUX.1-dev-Controlnet-Inpainting-Beta.py)|[code](./model_inference_low_vram/FLUX.1-dev-Controlnet-Inpainting-Beta.py)|[code](./model_training/full/FLUX.1-dev-Controlnet-Inpainting-Beta.sh)|[code](./model_training/validate_full/FLUX.1-dev-Controlnet-Inpainting-Beta.py)|[code](./model_training/lora/FLUX.1-dev-Controlnet-Inpainting-Beta.sh)|[code](./model_training/validate_lora/FLUX.1-dev-Controlnet-Inpainting-Beta.py)| |[FLUX.1-dev-Controlnet-Union-alpha](https://www.modelscope.cn/models/InstantX/FLUX.1-dev-Controlnet-Union-alpha)|`controlnet_inputs`|[code](./model_inference/FLUX.1-dev-Controlnet-Union-alpha.py)|[code](./model_inference_low_vram/FLUX.1-dev-Controlnet-Union-alpha.py)|[code](./model_training/full/FLUX.1-dev-Controlnet-Union-alpha.sh)|[code](./model_training/validate_full/FLUX.1-dev-Controlnet-Union-alpha.py)|[code](./model_training/lora/FLUX.1-dev-Controlnet-Union-alpha.sh)|[code](./model_training/validate_lora/FLUX.1-dev-Controlnet-Union-alpha.py)| diff --git a/examples/flux/model_inference/FLUX.1-Kera-dev.py b/examples/flux/model_inference/FLUX.1-Krea-dev.py similarity index 94% rename from examples/flux/model_inference/FLUX.1-Kera-dev.py rename to examples/flux/model_inference/FLUX.1-Krea-dev.py index af425d1..786ab32 100644 --- a/examples/flux/model_inference/FLUX.1-Kera-dev.py +++ b/examples/flux/model_inference/FLUX.1-Krea-dev.py @@ -17,11 +17,11 @@ prompt = "An beautiful woman is riding a bicycle in a park, wearing a red dress" negative_prompt = "worst quality, low quality, monochrome, zombie, interlocked fingers, Aissist, cleavage, nsfw," image = pipe(prompt=prompt, seed=0, embedded_guidance=4.5) -image.save("flux_kera.jpg") +image.save("flux_krea.jpg") image = pipe( prompt=prompt, negative_prompt=negative_prompt, seed=0, cfg_scale=2, num_inference_steps=50, embedded_guidance=4.5 ) -image.save("flux_kera_cfg.jpg") +image.save("flux_krea_cfg.jpg") diff --git a/examples/flux/model_inference_low_vram/FLUX.1-Kera-dev.py b/examples/flux/model_inference_low_vram/FLUX.1-Krea-dev.py similarity index 95% rename from examples/flux/model_inference_low_vram/FLUX.1-Kera-dev.py rename to examples/flux/model_inference_low_vram/FLUX.1-Krea-dev.py index 83c14e7..4cf946a 100644 --- a/examples/flux/model_inference_low_vram/FLUX.1-Kera-dev.py +++ b/examples/flux/model_inference_low_vram/FLUX.1-Krea-dev.py @@ -18,11 +18,11 @@ prompt = "An beautiful woman is riding a bicycle in a park, wearing a red dress" negative_prompt = "worst quality, low quality, monochrome, zombie, interlocked fingers, Aissist, cleavage, nsfw," image = pipe(prompt=prompt, seed=0, embedded_guidance=4.5) -image.save("flux_kera.jpg") +image.save("flux_krea.jpg") image = pipe( prompt=prompt, negative_prompt=negative_prompt, seed=0, cfg_scale=2, num_inference_steps=50, embedded_guidance=4.5 ) -image.save("flux_kera_cfg.jpg") +image.save("flux_krea_cfg.jpg") diff --git a/examples/flux/model_training/full/FLUX.1-Kera-dev.sh b/examples/flux/model_training/full/FLUX.1-Krea-dev.sh similarity index 92% rename from examples/flux/model_training/full/FLUX.1-Kera-dev.sh rename to examples/flux/model_training/full/FLUX.1-Krea-dev.sh index 3b3b409..76727a8 100644 --- a/examples/flux/model_training/full/FLUX.1-Kera-dev.sh +++ b/examples/flux/model_training/full/FLUX.1-Krea-dev.sh @@ -7,6 +7,6 @@ accelerate launch --config_file examples/flux/model_training/full/accelerate_con --learning_rate 1e-5 \ --num_epochs 1 \ --remove_prefix_in_ckpt "pipe.dit." \ - --output_path "./models/train/FLUX.1-Kera-dev_full" \ + --output_path "./models/train/FLUX.1-Krea-dev_full" \ --trainable_models "dit" \ --use_gradient_checkpointing diff --git a/examples/flux/model_training/lora/FLUX.1-Kera-dev.sh b/examples/flux/model_training/lora/FLUX.1-Krea-dev.sh similarity index 93% rename from examples/flux/model_training/lora/FLUX.1-Kera-dev.sh rename to examples/flux/model_training/lora/FLUX.1-Krea-dev.sh index 82a2b8c..1d54d43 100644 --- a/examples/flux/model_training/lora/FLUX.1-Kera-dev.sh +++ b/examples/flux/model_training/lora/FLUX.1-Krea-dev.sh @@ -7,7 +7,7 @@ accelerate launch examples/flux/model_training/train.py \ --learning_rate 1e-4 \ --num_epochs 5 \ --remove_prefix_in_ckpt "pipe.dit." \ - --output_path "./models/train/FLUX.1-Kera-dev_lora" \ + --output_path "./models/train/FLUX.1-Krea-dev_lora" \ --lora_base_model "dit" \ --lora_target_modules "a_to_qkv,b_to_qkv,ff_a.0,ff_a.2,ff_b.0,ff_b.2,a_to_out,b_to_out,proj_out,norm.linear,norm1_a.linear,norm1_b.linear,to_qkv_mlp" \ --lora_rank 32 \ diff --git a/examples/flux/model_training/validate_full/FLUX.1-Kera-dev.py b/examples/flux/model_training/validate_full/FLUX.1-Krea-dev.py similarity index 87% rename from examples/flux/model_training/validate_full/FLUX.1-Kera-dev.py rename to examples/flux/model_training/validate_full/FLUX.1-Krea-dev.py index f3b433d..bceb464 100644 --- a/examples/flux/model_training/validate_full/FLUX.1-Kera-dev.py +++ b/examples/flux/model_training/validate_full/FLUX.1-Krea-dev.py @@ -13,8 +13,8 @@ pipe = FluxImagePipeline.from_pretrained( ModelConfig(model_id="black-forest-labs/FLUX.1-dev", origin_file_pattern="ae.safetensors"), ], ) -state_dict = load_state_dict("models/train/FLUX.1-Kera-dev_full/epoch-0.safetensors") +state_dict = load_state_dict("models/train/FLUX.1-Krea-dev_full/epoch-0.safetensors") pipe.dit.load_state_dict(state_dict) image = pipe(prompt="a dog", seed=0) -image.save("image_FLUX.1-Kera-dev_full.jpg") +image.save("image_FLUX.1-Krea-dev_full.jpg") diff --git a/examples/flux/model_training/validate_lora/FLUX.1-Kera-dev.py b/examples/flux/model_training/validate_lora/FLUX.1-Krea-dev.py similarity index 86% rename from examples/flux/model_training/validate_lora/FLUX.1-Kera-dev.py rename to examples/flux/model_training/validate_lora/FLUX.1-Krea-dev.py index 6bda583..068ce72 100644 --- a/examples/flux/model_training/validate_lora/FLUX.1-Kera-dev.py +++ b/examples/flux/model_training/validate_lora/FLUX.1-Krea-dev.py @@ -12,7 +12,7 @@ pipe = FluxImagePipeline.from_pretrained( ModelConfig(model_id="black-forest-labs/FLUX.1-dev", origin_file_pattern="ae.safetensors"), ], ) -pipe.load_lora(pipe.dit, "models/train/FLUX.1-Kera-dev_lora/epoch-4.safetensors", alpha=1) +pipe.load_lora(pipe.dit, "models/train/FLUX.1-Krea-dev_lora/epoch-4.safetensors", alpha=1) image = pipe(prompt="a dog", seed=0) -image.save("image_FLUX.1-Kera-dev_lora.jpg") +image.save("image_FLUX.1-Krea-dev_lora.jpg")