Skip to content

bytedance/DreamO

Repository files navigation

DreamO

Official implementation of DreamO: A Unified Framework for Image Customization

arXiv demo

🚩 Updates

  • 2025.08.11: 🎉 DreamO accepted by SIGGRAPH Asia 2025!
  • 2025.06.24: 🔥🔥We are excited to release DreamO v1.1 with significant improvements in image quality, reduced likelihood of body composition errors, and enhanced aesthetics. Learn more about the model
  • 2025.06.26: Nunchaku is now supported for model quantization.
  • 2025.05.30: 🔥 Native ComfyUI implementation is now available!
  • 2025.05.12: Support consumer-grade GPUs (16GB or 24GB) now, see here for instruction
  • 2025.05.11: We have updated the model to mitigate over-saturation and plastic-face issue. The new version shows consistent improvements over the previous release. Please check it out!
  • 2025.05.08: release codes and models
  • 2025.04.24: release DreamO tech report.
output.mp4

🔧 Dependencies and Installation

note for v1.1: In order to use Nunchaku for model quantization, we have updated the diffusers version to 0.33.1. If you have the older version 0.31.0 installed, please update diffusers; otherwise, the code will throw errors.

# clone DreamO repo
git clone https://github.com/bytedance/DreamO.git
cd DreamO
# create conda env
conda create --name dreamo python=3.10
# activate env
conda activate dreamo
# install dependent packages
pip install -r requirements.txt

(optional) Nunchaku: If you want to use Nunchaku for model quantization, please refer to the original repo for installation guide.

⚡ Quick Inference

Local Gradio Demo

python app.py
options:
  --version {v1.1,v1}   default will use the latest v1.1 model, you can also switch back to v1
  --offload             Enable 'quant=nunchaku' and 'offload' to reduce the original 24GB VRAM to 6.5GB.
  --no_turbo            Use turbo to reduce the original 25 steps to 12 steps.
  --quant {none,int8,nunchaku}
                        Quantize to use: none(bf16), int8, nunchaku
  --device DEVICE       Device to use: auto, cuda, mps, or cpu

We observe strong compatibility between DreamO and the accelerated FLUX LoRA variant (FLUX-turbo), and thus enable Turbo LoRA by default, reducing inference to 12 steps (vs. 25+ by default). Turbo can be disabled via --no_turbo, though our evaluation shows mixed results; we therefore recommend keeping Turbo enabled.

tips: If you observe limb distortion or poor text generation, try increasing the guidance scale; if the image appears overly glossy or over-saturated, consider lowering the guidance scale.

For consumer-grade GPUs

Currently, the code supports two quantization schemes: int8 from optimum-quanto and Nunchaku. You can choose either one based on your needs and the actual results.

  • For users with 8GB GPUs, run python app.py --nunchaku --offload to enable CPU offloading alongside nunchaku quantization. According to the feedback, it takes about 20 seconds to generate a 1024-resolution image on NVIDIA 3080.

  • For users with 24GB GPUs, run python app.py --quant int8 to enable the int8-quantized model or python app.py --quant nunchaku to enable the nunchaku-quantized model.

  • For users with 16GB GPUs, run python app.py --int8 --offload to enable CPU offloading alongside int8 quantization. Note that CPU offload significantly reduces inference speed and should only be enabled when necessary.

For macOS Apple Silicon (M1/M2/M3/M4)

DreamO now supports macOS with Apple Silicon chips using Metal Performance Shaders (MPS). The app automatically detects and uses MPS when available.

  • For macOS users, simply run python app.py and the app will automatically use MPS acceleration.
  • Manual device selection: You can explicitly specify the device using python app.py --device mps (or --device cpu if needed).
  • Memory optimization: For devices with limited memory, you can combine MPS with quantization: python app.py --device mps --int8

Note: Make sure you have PyTorch with MPS support installed. The current requirements.txt includes PyTorch 2.6.0+ which has full MPS support.

Supported Tasks

IP

This task is similar to IP-Adapter and supports a wide range of inputs including characters, objects, and animals. By leveraging VAE-based feature encoding, DreamO achieves higher fidelity than previous adapter methods, with a distinct advantage in preserving character identity.

IP_example

ID

Here, ID specifically refers to facial identity. Unlike the IP task, which considers both face and clothing, the ID task focuses solely on facial features. This task is similar to InstantID and PuLID. Compared to previous methods, DreamO achieves higher facial fidelity, but introduces more model contamination than the SOTA approach PuLID.

ID_example

tips: If you notice the face appears overly glossy, try lowering the guidance scale.

Try-On

This task supports inputs such as tops, bottoms, glasses, and hats, and enables virtual try-on with multiple garments. Notably, our training set does not include multi-garment or ID+garment data, yet the model generalizes well to these unseen combinations.

tryon_example

Style

This task is similar to Style-Adapter and InstantStyle. Please note that style consistency is currently less stable compared to other tasks, and in the current version, style cannot be combined with other conditions. We are working on improvements in future releases—stay tuned.

style_example

Multi Condition

You can use multiple conditions (ID, IP, Try-On) to generate more creative images. Thanks to the feature routing constraint proposed in the paper, DreamO effectively mitigates conflicts and entanglement among multiple entities.

multi_cond_example

ComfyUI

Online HuggingFace Demo

You can try DreamO demo on HuggingFace.

Disclaimer

This project strives to impact the domain of AI-driven image generation positively. Users are granted the freedom to create images using this tool, but they are expected to comply with local laws and utilize it responsibly. The developers do not assume any responsibility for potential misuse by users.

Citation

If DreamO is helpful, please help to ⭐ the repo.

If you find this project useful for your research, please consider citing our paper.

📧 Contact

If you have any comments or questions, please open a new issue or contact Yanze Wu and Chong Mou.

About

[SIGGRAPH Asia 2025] DreamO: A Unified Framework for Image Customization

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages