Upload README.md
Browse files
README.md
CHANGED
@@ -1,7 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
<div align="center">
|
2 |
|
3 |
# ComfyUI
|
4 |
-
**The most powerful and modular
|
5 |
|
6 |
|
7 |
[![Website][website-shield]][website-url]
|
@@ -31,48 +43,17 @@
|
|
31 |

|
32 |
</div>
|
33 |
|
34 |
-
|
35 |
-
|
36 |
-
## Get Started
|
37 |
-
|
38 |
-
#### [Desktop Application](https://www.comfy.org/download)
|
39 |
-
- The easiest way to get started.
|
40 |
-
- Available on Windows & macOS.
|
41 |
-
|
42 |
-
#### [Windows Portable Package](#installing)
|
43 |
-
- Get the latest commits and completely portable.
|
44 |
-
- Available on Windows.
|
45 |
-
|
46 |
-
#### [Manual Install](#manual-install-windows-linux)
|
47 |
-
Supports all operating systems and GPU types (NVIDIA, AMD, Intel, Apple Silicon, Ascend).
|
48 |
-
|
49 |
-
## [Examples](https://comfyanonymous.github.io/ComfyUI_examples/)
|
50 |
-
See what ComfyUI can do with the [example workflows](https://comfyanonymous.github.io/ComfyUI_examples/).
|
51 |
|
|
|
52 |
|
53 |
## Features
|
54 |
- Nodes/graph/flowchart interface to experiment and create complex Stable Diffusion workflows without needing to code anything.
|
55 |
-
-
|
56 |
-
|
57 |
-
|
58 |
-
|
59 |
-
- [SD3 and SD3.5](https://comfyanonymous.github.io/ComfyUI_examples/sd3/)
|
60 |
-
- Pixart Alpha and Sigma
|
61 |
-
- [AuraFlow](https://comfyanonymous.github.io/ComfyUI_examples/aura_flow/)
|
62 |
-
- [HunyuanDiT](https://comfyanonymous.github.io/ComfyUI_examples/hunyuan_dit/)
|
63 |
-
- [Flux](https://comfyanonymous.github.io/ComfyUI_examples/flux/)
|
64 |
-
- [Lumina Image 2.0](https://comfyanonymous.github.io/ComfyUI_examples/lumina2/)
|
65 |
-
- [HiDream](https://comfyanonymous.github.io/ComfyUI_examples/hidream/)
|
66 |
-
- Video Models
|
67 |
-
- [Stable Video Diffusion](https://comfyanonymous.github.io/ComfyUI_examples/video/)
|
68 |
-
- [Mochi](https://comfyanonymous.github.io/ComfyUI_examples/mochi/)
|
69 |
-
- [LTX-Video](https://comfyanonymous.github.io/ComfyUI_examples/ltxv/)
|
70 |
-
- [Hunyuan Video](https://comfyanonymous.github.io/ComfyUI_examples/hunyuan_video/)
|
71 |
-
- [Nvidia Cosmos](https://comfyanonymous.github.io/ComfyUI_examples/cosmos/)
|
72 |
-
- [Wan 2.1](https://comfyanonymous.github.io/ComfyUI_examples/wan/)
|
73 |
-
- 3D Models
|
74 |
-
- [Hunyuan3D 2.0](https://docs.comfy.org/tutorials/3d/hunyuan3D-2)
|
75 |
-
- [Stable Audio](https://comfyanonymous.github.io/ComfyUI_examples/audio/)
|
76 |
- Asynchronous Queue system
|
77 |
- Many optimizations: Only re-executes the parts of the workflow that changes between executions.
|
78 |
- Smart memory management: can automatically run models on GPUs with as low as 1GB vram.
|
@@ -92,6 +73,9 @@ See what ComfyUI can do with the [example workflows](https://comfyanonymous.gith
|
|
92 |
- [GLIGEN](https://comfyanonymous.github.io/ComfyUI_examples/gligen/)
|
93 |
- [Model Merging](https://comfyanonymous.github.io/ComfyUI_examples/model_merging/)
|
94 |
- [LCM models and Loras](https://comfyanonymous.github.io/ComfyUI_examples/lcm/)
|
|
|
|
|
|
|
95 |
- Latent previews with [TAESD](#how-to-show-high-quality-previews)
|
96 |
- Starts up very fast.
|
97 |
- Works fully offline: will never download anything.
|
@@ -129,8 +113,6 @@ Workflow examples can be found on the [Examples page](https://comfyanonymous.git
|
|
129 |
| `Q` | Toggle visibility of the queue |
|
130 |
| `H` | Toggle visibility of history |
|
131 |
| `R` | Refresh graph |
|
132 |
-
| `F` | Show/Hide menu |
|
133 |
-
| `.` | Fit view to selection (Whole graph when nothing is selected) |
|
134 |
| Double-Click LMB | Open node quick search palette |
|
135 |
| `Shift` + Drag | Move multiple wires at once |
|
136 |
| `Ctrl` + `Alt` + LMB | Disconnect all wires from clicked slot |
|
@@ -139,7 +121,7 @@ Workflow examples can be found on the [Examples page](https://comfyanonymous.git
|
|
139 |
|
140 |
# Installing
|
141 |
|
142 |
-
## Windows
|
143 |
|
144 |
There is a portable standalone build for Windows that should work for running on Nvidia GPUs or for running on your CPU only on the [releases page](https://github.com/comfyanonymous/ComfyUI/releases).
|
145 |
|
@@ -149,8 +131,6 @@ Simply download, extract with [7-Zip](https://7-zip.org) and run. Make sure you
|
|
149 |
|
150 |
If you have trouble extracting it, right click the file -> properties -> unblock
|
151 |
|
152 |
-
If you have a 50 series Blackwell card like a 5090 or 5080 see [this discussion thread](https://github.com/comfyanonymous/ComfyUI/discussions/6643)
|
153 |
-
|
154 |
#### How do I share models between another UI and ComfyUI?
|
155 |
|
156 |
See the [Config file](extra_model_paths.yaml.example) to set the search paths for models. In the standalone windows build you can find this file in the ComfyUI directory. Rename this file to extra_model_paths.yaml and edit it with your favorite text editor.
|
@@ -159,18 +139,9 @@ See the [Config file](extra_model_paths.yaml.example) to set the search paths fo
|
|
159 |
|
160 |
To run it on services like paperspace, kaggle or colab you can use my [Jupyter Notebook](notebooks/comfyui_colab.ipynb)
|
161 |
|
162 |
-
|
163 |
-
## [comfy-cli](https://docs.comfy.org/comfy-cli/getting-started)
|
164 |
-
|
165 |
-
You can install and start ComfyUI using comfy-cli:
|
166 |
-
```bash
|
167 |
-
pip install comfy-cli
|
168 |
-
comfy install
|
169 |
-
```
|
170 |
-
|
171 |
## Manual Install (Windows, Linux)
|
172 |
|
173 |
-
python 3.13
|
174 |
|
175 |
Git clone this repo.
|
176 |
|
@@ -182,45 +153,21 @@ Put your VAE in: models/vae
|
|
182 |
### AMD GPUs (Linux only)
|
183 |
AMD users can install rocm and pytorch with pip if you don't have it already installed, this is the command to install the stable version:
|
184 |
|
185 |
-
```pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm6.2
|
186 |
-
|
187 |
-
This is the command to install the nightly with ROCm 6.3 which might have some performance improvements:
|
188 |
-
|
189 |
-
```pip install --pre torch torchvision torchaudio --index-url https://download.pytorch.org/whl/nightly/rocm6.3```
|
190 |
-
|
191 |
-
### Intel GPUs (Windows and Linux)
|
192 |
-
|
193 |
-
(Option 1) Intel Arc GPU users can install native PyTorch with torch.xpu support using pip (currently available in PyTorch nightly builds). More information can be found [here](https://pytorch.org/docs/main/notes/get_start_xpu.html)
|
194 |
-
|
195 |
-
1. To install PyTorch nightly, use the following command:
|
196 |
-
|
197 |
-
```pip install --pre torch torchvision torchaudio --index-url https://download.pytorch.org/whl/nightly/xpu```
|
198 |
-
|
199 |
-
2. Launch ComfyUI by running `python main.py`
|
200 |
-
|
201 |
-
|
202 |
-
(Option 2) Alternatively, Intel GPUs supported by Intel Extension for PyTorch (IPEX) can leverage IPEX for improved performance.
|
203 |
|
204 |
-
|
205 |
|
206 |
-
```
|
207 |
-
conda install libuv
|
208 |
-
pip install torch==2.3.1.post0+cxx11.abi torchvision==0.18.1.post0+cxx11.abi torchaudio==2.3.1.post0+cxx11.abi intel-extension-for-pytorch==2.3.110.post0+xpu --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/cn/
|
209 |
-
```
|
210 |
-
|
211 |
-
For other supported Intel GPUs with IPEX, visit [Installation](https://intel.github.io/intel-extension-for-pytorch/index.html#installation?platform=gpu) for more information.
|
212 |
-
|
213 |
-
Additional discussion and help can be found [here](https://github.com/comfyanonymous/ComfyUI/discussions/476).
|
214 |
|
215 |
### NVIDIA
|
216 |
|
217 |
Nvidia users should install stable pytorch using this command:
|
218 |
|
219 |
-
```pip install torch torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/
|
220 |
|
221 |
-
This is the command to install pytorch nightly instead which
|
222 |
|
223 |
-
```pip install --pre torch torchvision torchaudio --index-url https://download.pytorch.org/whl/nightly/
|
224 |
|
225 |
#### Troubleshooting
|
226 |
|
@@ -240,6 +187,17 @@ After this you should have everything installed and can proceed to running Comfy
|
|
240 |
|
241 |
### Others:
|
242 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
243 |
#### Apple Mac silicon
|
244 |
|
245 |
You can install ComfyUI in Apple Mac silicon (M1 or M2) with any recent macOS version.
|
@@ -255,23 +213,6 @@ You can install ComfyUI in Apple Mac silicon (M1 or M2) with any recent macOS ve
|
|
255 |
|
256 |
```pip install torch-directml``` Then you can launch ComfyUI with: ```python main.py --directml```
|
257 |
|
258 |
-
#### Ascend NPUs
|
259 |
-
|
260 |
-
For models compatible with Ascend Extension for PyTorch (torch_npu). To get started, ensure your environment meets the prerequisites outlined on the [installation](https://ascend.github.io/docs/sources/ascend/quick_install.html) page. Here's a step-by-step guide tailored to your platform and installation method:
|
261 |
-
|
262 |
-
1. Begin by installing the recommended or newer kernel version for Linux as specified in the Installation page of torch-npu, if necessary.
|
263 |
-
2. Proceed with the installation of Ascend Basekit, which includes the driver, firmware, and CANN, following the instructions provided for your specific platform.
|
264 |
-
3. Next, install the necessary packages for torch-npu by adhering to the platform-specific instructions on the [Installation](https://ascend.github.io/docs/sources/pytorch/install.html#pytorch) page.
|
265 |
-
4. Finally, adhere to the [ComfyUI manual installation](#manual-install-windows-linux) guide for Linux. Once all components are installed, you can run ComfyUI as described earlier.
|
266 |
-
|
267 |
-
#### Cambricon MLUs
|
268 |
-
|
269 |
-
For models compatible with Cambricon Extension for PyTorch (torch_mlu). Here's a step-by-step guide tailored to your platform and installation method:
|
270 |
-
|
271 |
-
1. Install the Cambricon CNToolkit by adhering to the platform-specific instructions on the [Installation](https://www.cambricon.com/docs/sdk_1.15.0/cntoolkit_3.7.2/cntoolkit_install_3.7.2/index.html)
|
272 |
-
2. Next, install the PyTorch(torch_mlu) following the instructions on the [Installation](https://www.cambricon.com/docs/sdk_1.15.0/cambricon_pytorch_1.17.0/user_guide_1.9/index.html)
|
273 |
-
3. Launch ComfyUI by running `python main.py`
|
274 |
-
|
275 |
# Running
|
276 |
|
277 |
```python main.py```
|
@@ -290,8 +231,6 @@ You can enable experimental memory efficient attention on pytorch 2.5 in ComfyUI
|
|
290 |
|
291 |
```TORCH_ROCM_AOTRITON_ENABLE_EXPERIMENTAL=1 python main.py --use-pytorch-cross-attention```
|
292 |
|
293 |
-
You can also try setting this env variable `PYTORCH_TUNABLEOP_ENABLED=1` which might speed things up at the cost of a very slow initial run.
|
294 |
-
|
295 |
# Notes
|
296 |
|
297 |
Only parts of the graph that have an output with all the correct inputs will be executed.
|
@@ -327,8 +266,6 @@ Use `--tls-keyfile key.pem --tls-certfile cert.pem` to enable TLS/SSL, the app w
|
|
327 |
|
328 |
## Support and dev channel
|
329 |
|
330 |
-
[Discord](https://comfy.org/discord): Try the #help or #feedback channels.
|
331 |
-
|
332 |
[Matrix space: #comfyui_space:matrix.org](https://app.element.io/#/room/%23comfyui_space%3Amatrix.org) (it's like discord but open source).
|
333 |
|
334 |
See also: [https://www.comfy.org/](https://www.comfy.org/)
|
@@ -345,7 +282,7 @@ For any bugs, issues, or feature requests related to the frontend, please use th
|
|
345 |
|
346 |
The new frontend is now the default for ComfyUI. However, please note:
|
347 |
|
348 |
-
1. The frontend in the main ComfyUI repository is updated
|
349 |
2. Daily releases are available in the separate frontend repository.
|
350 |
|
351 |
To use the most up-to-date frontend version:
|
@@ -362,7 +299,7 @@ To use the most up-to-date frontend version:
|
|
362 |
--front-end-version Comfy-Org/ComfyUI_frontend@1.2.2
|
363 |
```
|
364 |
|
365 |
-
This approach allows you to easily switch between the stable
|
366 |
|
367 |
### Accessing the Legacy Frontend
|
368 |
|
@@ -378,4 +315,4 @@ This will use a snapshot of the legacy frontend preserved in the [ComfyUI Legacy
|
|
378 |
|
379 |
### Which GPU should I buy for this?
|
380 |
|
381 |
-
[See this page for some recommendations](https://github.com/comfyanonymous/ComfyUI/wiki/Which-GPU-should-I-buy-for-ComfyUI)
|
|
|
1 |
+
---
|
2 |
+
title: Flux Style Shaping
|
3 |
+
emoji: 🚀
|
4 |
+
colorFrom: indigo
|
5 |
+
colorTo: gray
|
6 |
+
sdk: gradio
|
7 |
+
sdk_version: 5.12.0
|
8 |
+
app_file: app.py
|
9 |
+
pinned: false
|
10 |
+
license: mit
|
11 |
+
short_description: Optical illusions and style transfer with FLUX
|
12 |
+
---
|
13 |
<div align="center">
|
14 |
|
15 |
# ComfyUI
|
16 |
+
**The most powerful and modular diffusion model GUI and backend.**
|
17 |
|
18 |
|
19 |
[![Website][website-shield]][website-url]
|
|
|
43 |

|
44 |
</div>
|
45 |
|
46 |
+
This ui will let you design and execute advanced stable diffusion pipelines using a graph/nodes/flowchart based interface. For some workflow examples and see what ComfyUI can do you can check out:
|
47 |
+
### [ComfyUI Examples](https://comfyanonymous.github.io/ComfyUI_examples/)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
48 |
|
49 |
+
### [Installing ComfyUI](#installing)
|
50 |
|
51 |
## Features
|
52 |
- Nodes/graph/flowchart interface to experiment and create complex Stable Diffusion workflows without needing to code anything.
|
53 |
+
- Fully supports SD1.x, SD2.x, [SDXL](https://comfyanonymous.github.io/ComfyUI_examples/sdxl/), [Stable Video Diffusion](https://comfyanonymous.github.io/ComfyUI_examples/video/), [Stable Cascade](https://comfyanonymous.github.io/ComfyUI_examples/stable_cascade/), [SD3](https://comfyanonymous.github.io/ComfyUI_examples/sd3/) and [Stable Audio](https://comfyanonymous.github.io/ComfyUI_examples/audio/)
|
54 |
+
- [LTX-Video](https://comfyanonymous.github.io/ComfyUI_examples/ltxv/)
|
55 |
+
- [Flux](https://comfyanonymous.github.io/ComfyUI_examples/flux/)
|
56 |
+
- [Mochi](https://comfyanonymous.github.io/ComfyUI_examples/mochi/)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
57 |
- Asynchronous Queue system
|
58 |
- Many optimizations: Only re-executes the parts of the workflow that changes between executions.
|
59 |
- Smart memory management: can automatically run models on GPUs with as low as 1GB vram.
|
|
|
73 |
- [GLIGEN](https://comfyanonymous.github.io/ComfyUI_examples/gligen/)
|
74 |
- [Model Merging](https://comfyanonymous.github.io/ComfyUI_examples/model_merging/)
|
75 |
- [LCM models and Loras](https://comfyanonymous.github.io/ComfyUI_examples/lcm/)
|
76 |
+
- [SDXL Turbo](https://comfyanonymous.github.io/ComfyUI_examples/sdturbo/)
|
77 |
+
- [AuraFlow](https://comfyanonymous.github.io/ComfyUI_examples/aura_flow/)
|
78 |
+
- [HunyuanDiT](https://comfyanonymous.github.io/ComfyUI_examples/hunyuan_dit/)
|
79 |
- Latent previews with [TAESD](#how-to-show-high-quality-previews)
|
80 |
- Starts up very fast.
|
81 |
- Works fully offline: will never download anything.
|
|
|
113 |
| `Q` | Toggle visibility of the queue |
|
114 |
| `H` | Toggle visibility of history |
|
115 |
| `R` | Refresh graph |
|
|
|
|
|
116 |
| Double-Click LMB | Open node quick search palette |
|
117 |
| `Shift` + Drag | Move multiple wires at once |
|
118 |
| `Ctrl` + `Alt` + LMB | Disconnect all wires from clicked slot |
|
|
|
121 |
|
122 |
# Installing
|
123 |
|
124 |
+
## Windows
|
125 |
|
126 |
There is a portable standalone build for Windows that should work for running on Nvidia GPUs or for running on your CPU only on the [releases page](https://github.com/comfyanonymous/ComfyUI/releases).
|
127 |
|
|
|
131 |
|
132 |
If you have trouble extracting it, right click the file -> properties -> unblock
|
133 |
|
|
|
|
|
134 |
#### How do I share models between another UI and ComfyUI?
|
135 |
|
136 |
See the [Config file](extra_model_paths.yaml.example) to set the search paths for models. In the standalone windows build you can find this file in the ComfyUI directory. Rename this file to extra_model_paths.yaml and edit it with your favorite text editor.
|
|
|
139 |
|
140 |
To run it on services like paperspace, kaggle or colab you can use my [Jupyter Notebook](notebooks/comfyui_colab.ipynb)
|
141 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
142 |
## Manual Install (Windows, Linux)
|
143 |
|
144 |
+
Note that some dependencies do not yet support python 3.13 so using 3.12 is recommended.
|
145 |
|
146 |
Git clone this repo.
|
147 |
|
|
|
153 |
### AMD GPUs (Linux only)
|
154 |
AMD users can install rocm and pytorch with pip if you don't have it already installed, this is the command to install the stable version:
|
155 |
|
156 |
+
```pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm6.2```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
157 |
|
158 |
+
This is the command to install the nightly with ROCm 6.2 which might have some performance improvements:
|
159 |
|
160 |
+
```pip install --pre torch torchvision torchaudio --index-url https://download.pytorch.org/whl/nightly/rocm6.2```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
161 |
|
162 |
### NVIDIA
|
163 |
|
164 |
Nvidia users should install stable pytorch using this command:
|
165 |
|
166 |
+
```pip install torch torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/cu124```
|
167 |
|
168 |
+
This is the command to install pytorch nightly instead which might have performance improvements:
|
169 |
|
170 |
+
```pip install --pre torch torchvision torchaudio --index-url https://download.pytorch.org/whl/nightly/cu124```
|
171 |
|
172 |
#### Troubleshooting
|
173 |
|
|
|
187 |
|
188 |
### Others:
|
189 |
|
190 |
+
#### Intel GPUs
|
191 |
+
|
192 |
+
Intel GPU support is available for all Intel GPUs supported by Intel's Extension for Pytorch (IPEX) with the support requirements listed in the [Installation](https://intel.github.io/intel-extension-for-pytorch/index.html#installation?platform=gpu) page. Choose your platform and method of install and follow the instructions. The steps are as follows:
|
193 |
+
|
194 |
+
1. Start by installing the drivers or kernel listed or newer in the Installation page of IPEX linked above for Windows and Linux if needed.
|
195 |
+
1. Follow the instructions to install [Intel's oneAPI Basekit](https://www.intel.com/content/www/us/en/developer/tools/oneapi/base-toolkit-download.html) for your platform.
|
196 |
+
1. Install the packages for IPEX using the instructions provided in the Installation page for your platform.
|
197 |
+
1. Follow the [ComfyUI manual installation](#manual-install-windows-linux) instructions for Windows and Linux and run ComfyUI normally as described above after everything is installed.
|
198 |
+
|
199 |
+
Additional discussion and help can be found [here](https://github.com/comfyanonymous/ComfyUI/discussions/476).
|
200 |
+
|
201 |
#### Apple Mac silicon
|
202 |
|
203 |
You can install ComfyUI in Apple Mac silicon (M1 or M2) with any recent macOS version.
|
|
|
213 |
|
214 |
```pip install torch-directml``` Then you can launch ComfyUI with: ```python main.py --directml```
|
215 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
216 |
# Running
|
217 |
|
218 |
```python main.py```
|
|
|
231 |
|
232 |
```TORCH_ROCM_AOTRITON_ENABLE_EXPERIMENTAL=1 python main.py --use-pytorch-cross-attention```
|
233 |
|
|
|
|
|
234 |
# Notes
|
235 |
|
236 |
Only parts of the graph that have an output with all the correct inputs will be executed.
|
|
|
266 |
|
267 |
## Support and dev channel
|
268 |
|
|
|
|
|
269 |
[Matrix space: #comfyui_space:matrix.org](https://app.element.io/#/room/%23comfyui_space%3Amatrix.org) (it's like discord but open source).
|
270 |
|
271 |
See also: [https://www.comfy.org/](https://www.comfy.org/)
|
|
|
282 |
|
283 |
The new frontend is now the default for ComfyUI. However, please note:
|
284 |
|
285 |
+
1. The frontend in the main ComfyUI repository is updated weekly.
|
286 |
2. Daily releases are available in the separate frontend repository.
|
287 |
|
288 |
To use the most up-to-date frontend version:
|
|
|
299 |
--front-end-version Comfy-Org/ComfyUI_frontend@1.2.2
|
300 |
```
|
301 |
|
302 |
+
This approach allows you to easily switch between the stable weekly release and the cutting-edge daily updates, or even specific versions for testing purposes.
|
303 |
|
304 |
### Accessing the Legacy Frontend
|
305 |
|
|
|
315 |
|
316 |
### Which GPU should I buy for this?
|
317 |
|
318 |
+
[See this page for some recommendations](https://github.com/comfyanonymous/ComfyUI/wiki/Which-GPU-should-I-buy-for-ComfyUI)
|