From 6b8e3880ffcb3d301a9ab14e7feb57fe26b3d7f4 Mon Sep 17 00:00:00 2001 From: mi804 <1576993271@qq.com> Date: Fri, 1 Aug 2025 17:25:50 +0800 Subject: [PATCH] fix lowvram inference --- examples/flux/model_inference_low_vram/FLUX.1-Krea-dev.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/examples/flux/model_inference_low_vram/FLUX.1-Krea-dev.py b/examples/flux/model_inference_low_vram/FLUX.1-Krea-dev.py index 4cf946a..96f8414 100644 --- a/examples/flux/model_inference_low_vram/FLUX.1-Krea-dev.py +++ b/examples/flux/model_inference_low_vram/FLUX.1-Krea-dev.py @@ -6,7 +6,7 @@ pipe = FluxImagePipeline.from_pretrained( torch_dtype=torch.bfloat16, device="cuda", model_configs=[ - ModelConfig(model_id="black-forest-labs/FLUX.1-Krea-dev", origin_file_pattern="flux1-krea-dev.safetensors"), + ModelConfig(model_id="black-forest-labs/FLUX.1-Krea-dev", origin_file_pattern="flux1-krea-dev.safetensors", offload_device="cpu", offload_dtype=torch.float8_e4m3fn), ModelConfig(model_id="black-forest-labs/FLUX.1-dev", origin_file_pattern="text_encoder/model.safetensors", offload_device="cpu", offload_dtype=torch.float8_e4m3fn), ModelConfig(model_id="black-forest-labs/FLUX.1-dev", origin_file_pattern="text_encoder_2/", offload_device="cpu", offload_dtype=torch.float8_e4m3fn), ModelConfig(model_id="black-forest-labs/FLUX.1-dev", origin_file_pattern="ae.safetensors", offload_device="cpu", offload_dtype=torch.float8_e4m3fn),