From 1e58e6ef8205c93addd052b68a21b745a6d2ae49 Mon Sep 17 00:00:00 2001 From: feifeibear Date: Mon, 17 Mar 2025 09:00:52 +0000 Subject: [PATCH 1/2] fix some bugs --- diffsynth/distributed/__init__.py | 0 .../xdit_context_parallel.cpython-310.pyc | Bin 3754 -> 0 bytes requirements.txt | 1 + 3 files changed, 1 insertion(+) create mode 100644 diffsynth/distributed/__init__.py delete mode 100644 diffsynth/distributed/__pycache__/xdit_context_parallel.cpython-310.pyc diff --git a/diffsynth/distributed/__init__.py b/diffsynth/distributed/__init__.py new file mode 100644 index 0000000..e69de29 diff --git a/diffsynth/distributed/__pycache__/xdit_context_parallel.cpython-310.pyc b/diffsynth/distributed/__pycache__/xdit_context_parallel.cpython-310.pyc deleted file mode 100644 index 844c90076f202adc0be23a9f0c4326ec2fff0bdb..0000000000000000000000000000000000000000 GIT binary patch literal 0 HcmV?d00001 literal 3754 zcmb7HTW=f372cV>aJeEWN|r54b{wGyTC-^6I!M!7;N}JuMGN)81==apLHpwN5A=b2DUkk@KJ~4@f@jrpzev>q`-7-_Xp?0%LDV25%XRV)PQTfi>+Rn2h?Lk_QqN1D7 ztoz#TMI}g9XuF70&9MYoRmV{+;_>sWh)EdQ$MB@cVxigbrFr;0>l;NAnfs_~@vi80CC`N@b0#kL@XhIdv*l@vGL>vMr{-*?2$Y5&<~|m9u&Z#ESG2bx5vwLh$J1^xjB=^Sv9yz=vM{nT z(o|Q7#n-EuiU%>wHiQ)wnd$~*q{!wHnd%zM)>L?LTl3;Ah&MX3%NA)Ry9G z6(jBixj~!PXo ztW!RPqf}mHZ$Qez-n6Lp2?zR|(?0E3!;S&8Q`@kx-wIa6nw2vneJz}!w*O?~-cGS| zGu;uLoj337Py=7B^nMD5kf57#8=m8s%vzi>*-`14rL^YZcv1d~b0M zzl_JSC}pl|XHKRYy(G@VzKqI=lA4|9ZQvc?FA_0eF6<3u?;y`&AYMAq-}w5mFzXYP zh#m=d`HOv;h6|=5V5GvT8lUhjYu2n9 zSdYP&oLQ@CRL!bIbuuZz3}17>2N!654WDQDT1!4~nl&!2H?W0Er=G3S9x(LQ2Am2z z*|RpVd=0l)6*b{Mw5DyGbG>5arBmWSd?+_gEto$^?ea&~@Ajr^s9jP2t~`{Q`X~4| zM^|TCbL%gg%tC8>bZvC4YFBHg+Nm|VQIRy4`gt8=My7q6I4X)*EBnS1AA7I~ z1Dp$={0=qAc6x-U%ZI`mdFUmSq7Cv#t4|y7hyK}e_I?lD=MPP$Nt-AIE%ztb+@CG% z%l3^AA*OqAnor6CsNjt=F-hbFL*xLy+?#%rMg4JfcNFETy78J1%Xb3IsC!y|96%> z+`37=t0bHn*iv?Obc?{?WiT~iQ~LxtxaC-{n&Dh!BiS<%zCKBp*k)pXi?X4Y^Y!Ww zuo;h|feej+4WOF1r)%+eK%5lN%HA;UpQzioigpM(wI`4BC>3g#L}2jN-*Y?%hb zcqROkXsDR*|5sGt;Z)$ebG(GcGoUmud(j3`A^_8fTT=2w2=1NQNDo|u>SYU|zp{ag z7&B>vjW+Q9yiMcog|~+_pu{<+TB+joKcaukovnw`qhcH z59EndPt!|%Qcj_NI1!z2QfC93a&XqI(6^&4tZ1Yany zPqiVDn21r|OC7dJhlu^@5+D*V6+0kp)`sd@yoE2l4qxP9oOw?JFW~9`Ovotr_x3vL zi#)_bb`A?yggtPOIL-3H98qr`_vs(WZZA`E_d@Qx$ViO082i3)kuSmjm#|kJ)A<(v zUX-~irOGCYF@n(EuS`;^oqJi97~?QUyKvK&*Y)3Bfae{8+=AT02R1d>25ZnflEJkZ2rtAdv*C2+~ZGpE%?0g!G8f^oyg<> diff --git a/requirements.txt b/requirements.txt index 63a871b..7dc3846 100644 --- a/requirements.txt +++ b/requirements.txt @@ -11,3 +11,4 @@ sentencepiece protobuf modelscope ftfy +xfuser>=0.4.2 From d8b250607aac7764a5230c208b593ea5ae020c32 Mon Sep 17 00:00:00 2001 From: feifeibear Date: Mon, 17 Mar 2025 09:04:51 +0000 Subject: [PATCH 2/2] polish code --- diffsynth/models/wan_video_dit.py | 1 - examples/wanvideo/wan_14b_text_to_video.py | 23 ++++++++++++---------- 2 files changed, 13 insertions(+), 11 deletions(-) diff --git a/diffsynth/models/wan_video_dit.py b/diffsynth/models/wan_video_dit.py index 618e7d8..a2c55e1 100644 --- a/diffsynth/models/wan_video_dit.py +++ b/diffsynth/models/wan_video_dit.py @@ -91,7 +91,6 @@ def rope_apply(x, freqs, num_heads): x_out = torch.view_as_complex(x.to(torch.float64).reshape( x.shape[0], x.shape[1], x.shape[2], -1, 2)) - print(f"x_out.shape: {x_out.shape}, freqs.shape: {freqs.shape}") x_out = torch.view_as_real(x_out * freqs).flatten(2) return x_out.to(x.dtype) diff --git a/examples/wanvideo/wan_14b_text_to_video.py b/examples/wanvideo/wan_14b_text_to_video.py index d67e1d5..dcb2f29 100644 --- a/examples/wanvideo/wan_14b_text_to_video.py +++ b/examples/wanvideo/wan_14b_text_to_video.py @@ -5,22 +5,22 @@ import torch.distributed as dist # Download models -# snapshot_download("Wan-AI/Wan2.1-T2V-14B", local_dir="models/Wan-AI/Wan2.1-T2V-14B") +snapshot_download("Wan-AI/Wan2.1-T2V-14B", local_dir="models/Wan-AI/Wan2.1-T2V-14B") # Load models model_manager = ModelManager(device="cpu") model_manager.load_models( [ [ - "/demo-huabei2/models/dit/Wan2.1-T2V-14B/diffusion_pytorch_model-00001-of-00006.safetensors", - "/demo-huabei2/models/dit/Wan2.1-T2V-14B/diffusion_pytorch_model-00002-of-00006.safetensors", - "/demo-huabei2/models/dit/Wan2.1-T2V-14B/diffusion_pytorch_model-00003-of-00006.safetensors", - "/demo-huabei2/models/dit/Wan2.1-T2V-14B/diffusion_pytorch_model-00004-of-00006.safetensors", - "/demo-huabei2/models/dit/Wan2.1-T2V-14B/diffusion_pytorch_model-00005-of-00006.safetensors", - "/demo-huabei2/models/dit/Wan2.1-T2V-14B/diffusion_pytorch_model-00006-of-00006.safetensors", + "models/Wan-AI/Wan2.1-T2V-14B/diffusion_pytorch_model-00001-of-00006.safetensors", + "models/Wan-AI/Wan2.1-T2V-14B/diffusion_pytorch_model-00002-of-00006.safetensors", + "models/Wan-AI/Wan2.1-T2V-14B/diffusion_pytorch_model-00003-of-00006.safetensors", + "models/Wan-AI/Wan2.1-T2V-14B/diffusion_pytorch_model-00004-of-00006.safetensors", + "models/Wan-AI/Wan2.1-T2V-14B/diffusion_pytorch_model-00005-of-00006.safetensors", + "models/Wan-AI/Wan2.1-T2V-14B/diffusion_pytorch_model-00006-of-00006.safetensors", ], - "/demo-huabei2/models/dit/Wan2.1-T2V-14B/models_t5_umt5-xxl-enc-bf16.pth", - "/demo-huabei2/models/dit/Wan2.1-T2V-14B/Wan2.1_VAE.pth", + "models/Wan-AI/Wan2.1-T2V-14B/models_t5_umt5-xxl-enc-bf16.pth", + "models/Wan-AI/Wan2.1-T2V-14B/Wan2.1_VAE.pth", ], torch_dtype=torch.float8_e4m3fn, # You can set `torch_dtype=torch.bfloat16` to disable FP8 quantization. ) @@ -41,7 +41,10 @@ initialize_model_parallel( ) torch.cuda.set_device(dist.get_rank()) -pipe = WanVideoPipeline.from_model_manager(model_manager, torch_dtype=torch.bfloat16, device=f"cuda:{dist.get_rank()}", use_usp=True if dist.get_world_size() > 1 else False) +pipe = WanVideoPipeline.from_model_manager(model_manager, + torch_dtype=torch.bfloat16, + device=f"cuda:{dist.get_rank()}", + use_usp=True if dist.get_world_size() > 1 else False) pipe.enable_vram_management(num_persistent_param_in_dit=None) # You can set `num_persistent_param_in_dit` to a small number to reduce VRAM required. # Text-to-video