Skip to content

Commit bf45cdf

Browse files
committed
Support HiDream-I1-Dev and Lumina-Image-2.0, add cache_device option
1 parent 9db9e33 commit bf45cdf

15 files changed

Lines changed: 1794 additions & 541 deletions

README.md

Lines changed: 17 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,6 +6,12 @@ Timestep Embedding Aware Cache ([TeaCache](https://github.com/ali-vilab/TeaCache
66
TeaCache has now been integrated into ComfyUI and is compatible with the ComfyUI native nodes. ComfyUI-TeaCache is easy to use, simply connect the TeaCache node with the ComfyUI native nodes for seamless usage.
77

88
## Updates
9+
- Jun 15 2025: ComfyUI-TeaCache supports HiDream-I1-Dev and Lumina-Image-2.0, adds cache_device option:
10+
- It can achieve a 1.5x lossless speedup and a 2x speedup without much visual quality degradation for HiDream-I1-Dev.
11+
- Support HiDream-I1-Dev LoRA!
12+
- It can achieve a 1.5x lossless speedup and a 1.7x speedup without much visual quality degradation for Lumina-Image-2.0.
13+
- Support Lumina-Image-2.0 LoRA!
14+
- Add cache_device option according to the feedback from [3](https://github.com/welltop-cn/ComfyUI-TeaCache/issues/74), [4](https://github.com/welltop-cn/ComfyUI-TeaCache/issues/104) and [5](https://github.com/welltop-cn/ComfyUI-TeaCache/issues/143).
915
- May 22 2025: ComfyUI-TeaCache supports HiDream-I1-Full and redesigns TeaCache options:
1016
- It can achieve a 1.5x lossless speedup and a 2x speedup without much visual quality degradation.
1117
- Support HiDream-I1-Full LoRA!
@@ -62,7 +68,9 @@ To use TeaCache node, simply add `TeaCache` node to your workflow after `Load Di
6268
|:----------------------------:|:-----------------:|:-----------------:|:-----------------:|:-----------------:|
6369
| FLUX | 0.4 | 0 | 1 | ~2x |
6470
| PuLID-FLUX | 0.4 | 0 | 1 | ~1.7x |
71+
| HiDream-I1-Dev | 1 | 0 | 1 | ~2x |
6572
| HiDream-I1-Full | 0.35 | 0.1 | 1 | ~2x |
73+
| Lumina-Image-2.0 | 0.38 | 0.2 | 1 | ~1.7x |
6674
| HunyuanVideo | 0.15 | 0 | 1 | ~1.9x |
6775
| LTX-Video | 0.06 | 0 | 1 | ~1.7x |
6876
| CogVideoX | 0.3 | 0 | 1 | ~2x |
@@ -79,7 +87,9 @@ To use TeaCache node, simply add `TeaCache` node to your workflow after `Load Di
7987

8088
If the image/video after applying TeaCache is of low quality, please reduce rel_l1_thresh. I really don't recommend adjusting start_percent and end_percent unless you are an experienced engineer or creator.
8189

82-
The demo workflows ([flux](./examples/flux.json), [pulid_flux](./examples/pulid_flux.json), [hidream_i1_full](./examples/hidream_i1_full.json), [hunyuanvideo](./examples/hunyuanvideo.json), [ltx_video](./examples/ltx_video.json), [cogvideox](./examples/cogvideox.json), [wan2.1_t2v](./examples/wan2.1_t2v.json) and [wan2.1_i2v](./examples/wan2.1_i2v.json)) are placed in examples folder.
90+
If you have enough VRAM, please select `cuda` in the `cache_device` option, which can bring faster inference, but will increase VRAM slightly. If you have limited VRAM, please select `cpu` in the `cache_device` option, which do not increase VRAM, but will make inference slower slightly.
91+
92+
The demo workflows ([flux](./examples/flux.json), [pulid_flux](./examples/pulid_flux.json), [hidream_i1_dev](./examples/hidream_i1_dev.json), [hidream_i1_full](./examples/hidream_i1_full.json), [lumina_image_2](./examples/lumina_image_2.json), [hunyuanvideo](./examples/hunyuanvideo.json), [ltx_video](./examples/ltx_video.json), [cogvideox](./examples/cogvideox.json), [wan2.1_t2v](./examples/wan2.1_t2v.json) and [wan2.1_i2v](./examples/wan2.1_i2v.json)) are placed in examples folder.
8393

8494
### Compile Model
8595
To use Compile Model node, simply add `Compile Model` node to your workflow after `Load Diffusion Model` node or `TeaCache` node. Compile Model uses `torch.compile` to enhance the model performance by compiling model into more efficient intermediate representations (IRs). This compilation process leverages backend compilers to generate optimized code, which can significantly speed up inference. The compilation may take long time when you run the workflow at first, but once it is compiled, inference is extremely fast. The usage is shown below:
@@ -92,9 +102,15 @@ To use Compile Model node, simply add `Compile Model` node to your workflow afte
92102
- <p><strong>PuLID-FLUX</strong></p>
93103
![](./assets/compare_pulid_flux.png)
94104

105+
- <p><strong>HiDream-I1-Dev</strong></p>
106+
![](./assets/compare_hidream_i1_dev.png)
107+
95108
- <p><strong>HiDream-I1-Full</strong></p>
96109
![](./assets/compare_hidream_i1_full.png)
97110

111+
- <p><strong>Lumina-Image-2.0</strong></p>
112+
![](./assets/compare_lumina_image_2.png)
113+
98114
- <p><strong>HunyuanVideo</strong></p>
99115
https://github.com/user-attachments/assets/b3aca64d-c2ae-440c-a362-f3a7b6c633e0
100116

assets/compare_hidream_i1_dev.png

913 KB
Loading

assets/compare_lumina_image_2.png

902 KB
Loading

assets/compile.png

-13.1 KB
Loading

0 commit comments

Comments
 (0)