DiffSynth Studio
Introduction
DiffSynth Studio is a Diffusion engine. We have restructured architectures including Text Encoder, UNet, VAE, among others, maintaining compatibility with models from the open-source community while enhancing computational performance. We provide many interesting features. Enjoy the magic of Diffusion models!
Roadmap
- Aug 29, 2023. I propose DiffSynth, a video synthesis framework.
- Project Page.
- The source codes are released in EasyNLP.
- The technical report (ECML PKDD 2024) is released on arXiv.
- Oct 1, 2023. I release an early version of this project, namely FastSDXL. A try for building a diffusion engine.
- The source codes are released on GitHub.
- FastSDXL includes a trainable OLSS scheduler for efficiency improvement.
- Nov 15, 2023. I propose FastBlend, a powerful video deflickering algorithm.
- Dec 8, 2023. I decide to develop a new Project, aiming to release the potential of diffusion models, especially in video synthesis.
- Jan 29, 2024. I propose Diffutoon, a fantastic solution for toon shading.
- Project Page.
- The source codes are released in this project.
- The technical report (IJCAI 2024) is released on arXiv.
- Until now, DiffSynth Studio has supported the following models:
Installation
Create Python environment:
conda env create -f environment.yml
We find that sometimes conda cannot install cupy correctly, please install it manually. See this document for more details.
Enter the Python environment:
conda activate DiffSynthStudio
Usage (in WebUI)
python -m streamlit run DiffSynth_Studio.py
https://github.com/Artiprocher/DiffSynth-Studio/assets/35051019/93085557-73f3-4eee-a205-9829591ef954
Usage (in Python code)
The Python examples are in examples. We provide an overview here.
Image Synthesis
Generate high-resolution images, by breaking the limitation of diffusion models! examples/image_synthesis
| 512*512 | 1024*1024 | 2048*2048 | 4096*4096 |
|---|---|---|---|
| 1024*1024 | 2048*2048 |
|---|---|
Toon Shading
Render realistic videos in a flatten style and enable video editing features. examples/Diffutoon
https://github.com/Artiprocher/DiffSynth-Studio/assets/35051019/b54c05c5-d747-4709-be5e-b39af82404dd
https://github.com/Artiprocher/DiffSynth-Studio/assets/35051019/20528af5-5100-474a-8cdc-440b9efdd86c
Video Stylization
Video stylization without video models. examples/diffsynth
https://github.com/Artiprocher/DiffSynth-Studio/assets/35051019/59fb2f7b-8de0-4481-b79f-0c3a7361a1ea
Chinese Models
Use Hunyuan-DiT to generate images with Chinese prompts. We also support LoRA fine-tuning of this model. examples/hunyuan_dit
Prompt: 少女手捧鲜花,坐在公园的长椅上,夕阳的余晖洒在少女的脸庞,整个画面充满诗意的美感
| 1024x1024 | 2048x2048 (highres-fix) |
|---|---|
Prompt: 一只小狗蹦蹦跳跳,周围是姹紫嫣红的鲜花,远处是山脉
| Without LoRA | With LoRA |
|---|---|