diff --git a/docs/en/Pipeline_Usage/GPU_support.md b/docs/en/Pipeline_Usage/GPU_support.md index d6ecdd0..6c27de7 100644 --- a/docs/en/Pipeline_Usage/GPU_support.md +++ b/docs/en/Pipeline_Usage/GPU_support.md @@ -59,7 +59,7 @@ save_video(video, "video.mp4", fps=15, quality=5) ``` ### Training -NPU startup script samples have been added for each type of model,the scripts are stored in the `examples/xxx/special/npu_scripts`, for examples `examples/wanvideo/model_training/special/npu_scripts/Wan2.2-T2V-A14B-NPU.sh`. +NPU startup script samples have been added for each type of model,the scripts are stored in the `examples/xxx/special/npu_scripts`, for example `examples/wanvideo/model_training/special/npu_scripts/Wan2.2-T2V-A14B-NPU.sh`. In the NPU training scripts, NPU specific environment variables that can optimize performance have been added, and relevant parameters have been enabled for specific models. @@ -79,6 +79,6 @@ Set 0 or not set: indicates not enabling the binding function 2: Indicates enabling fine-grained kernel binding #### Parameters for specific models -| Model | Parameter | Note | -|-----------|---------------------------|-------------------| -| Wan 14B系列 | --initialize_model_on_cpu | The 14B model needs to be initialized on the CPU | \ No newline at end of file +| Model | Parameter | Note | +|----------------|---------------------------|-------------------| +| Wan 14B series | --initialize_model_on_cpu | The 14B model needs to be initialized on the CPU | \ No newline at end of file diff --git a/examples/qwen_image/model_training/special/npu_scripts/Qwen-Image-Edit-2509-LoRA-NPU.sh b/examples/qwen_image/model_training/special/npu_scripts/Qwen-Image-Edit-2509-LoRA-NPU.sh index 6599f51..9c3f02c 100644 --- a/examples/qwen_image/model_training/special/npu_scripts/Qwen-Image-Edit-2509-LoRA-NPU.sh +++ b/examples/qwen_image/model_training/special/npu_scripts/Qwen-Image-Edit-2509-LoRA-NPU.sh @@ -21,7 +21,7 @@ accelerate launch examples/qwen_image/model_training/train.py \ --task "sft:data_process" accelerate launch examples/qwen_image/model_training/train.py \ - --dataset_base_path "./models/train/Qwen-Image-LoRA-splited-cache" \ + --dataset_base_path "./models/train/Qwen-Image-Edit-2509-LoRA-splited-cache" \ --max_pixels 1048576 \ --dataset_repeat 50 \ --model_id_with_origin_paths "Qwen/Qwen-Image-Edit-2509:transformer/diffusion_pytorch_model*.safetensors" \ diff --git a/examples/z_image/model_training/special/npu_scripts/Z-Image-Turbo-NPU.sh b/examples/z_image/model_training/special/npu_scripts/Z-Image-Turbo-NPU.sh index ee9d6d6..93cc645 100644 --- a/examples/z_image/model_training/special/npu_scripts/Z-Image-Turbo-NPU.sh +++ b/examples/z_image/model_training/special/npu_scripts/Z-Image-Turbo-NPU.sh @@ -1,4 +1,3 @@ -# This example is tested on 8*A100 export PYTORCH_NPU_ALLOC_CONF=expandable_segments:True export CPU_AFFINITY_CONF=1