Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
@@ -1,269 +1,9 @@
|
|
1 |
-
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
|
8 |
-
5
|
9 |
-
|
10 |
-
5/18/2025
|
11 |
-
Add video extension tab. Currently only works with f1 model. Full credit to @pfqt and @chaojie for their amazing work!
|
12 |
-
|
13 |
-
# H1111
|
14 |
-
|
15 |
-
This is a GUI for tech wizard kohya-ss's musubi tuner's inference script.
|
16 |
-
https://github.com/kohya-ss/musubi-tuner
|
17 |
-
|
18 |
-
It allows inference with these models:
|
19 |
-
FramePack
|
20 |
-
Hunyuan-t2v
|
21 |
-
Hunyuan-i2v
|
22 |
-
Hunyuan-v2v
|
23 |
-
WanX-t2v
|
24 |
-
WanX-i2v
|
25 |
-
WanX-v2v
|
26 |
-
SkyReels-i2v
|
27 |
-
SkyReels-t2v
|
28 |
-
|
29 |
-
I have mostly been workiing on the framepack tab and the WanX-i2v tab. They are the best to use right now. WanX-i2v is used for skyreels v2 and the fun control models.
|
30 |
-
|
31 |
-
This supports queuing multiple different jobs if you open 2+ browser tabs and use the same model.
|
32 |
-
|
33 |
-
If you are running out of vram use more block swapping. Using FP8 scaled is also a decent option to lower memory usage, select fp8 and fp8 scaled to use it. Scaled fp8 tries to duplicate the important parts of the model from FP16. Sage attention is the fastest/lowest vram but difficult to install in windows.
|
34 |
-
|
35 |
-
Best quality will be obtained with only enabling block swapping and using the fp16 model with sdpa attention. You can speed things up with cfg skip, fp8 scaled, slg skip is small speedup, sage attention is fastest but all speedups come with quality degradations. I designed this to try to focus on quality over speed.
|
36 |
-
|
37 |
-
If you are using a lora that you didn't train with musubi you need to drag it to the convert lora tab and convert it to the default format. It should spit it out into the /lora folder.
|
38 |
-
|
39 |
-
If you need additional installation instructions or information create an issue and I will try to help. Also there are alot of settings notes on the musubi github linked above.
|
40 |
-
|
41 |
-
For torch 2.7.0 and windows installation try:
|
42 |
-
pip install typing-extensions
|
43 |
-
pip install torch==2.7.0+cu128 torchvision==0.22.0+cu128 --index-url https://download.pytorch.org/whl/cu128
|
44 |
-
pip install -r requirementsTorch27.txt
|
45 |
-
|
46 |
-
## To Use FramePack
|
47 |
-
|
48 |
-
|
49 |
-
download these 5 files from https://huggingface.co/maybleMyers/framepack_h1111 and put them in a subfolder named hunyuan (H1111/hunyuan), or reference where they are in the gui if you have already aquired them.
|
50 |
-
|
51 |
-
FramePackI2V_HY_bf16.safetensors or FramePack_F1_I2V_HY_20250503.safetensors for F1
|
52 |
-
|
53 |
-
clip_l.safetensors
|
54 |
-
|
55 |
-
llava_llama3_fp16.safetensors
|
56 |
-
|
57 |
-
model.safetensors
|
58 |
-
|
59 |
-
pytorch_model.pt
|
60 |
-
|
61 |
-
Lora trained with musubi tuner's framepack training confirmed to work great. Normal lora trained for hunyuan kinda suck. Use a lot of block swap this is a different back end than the official repo. If you select fp8 and fp8 scaled it will all fit on a 24gb gpu for fastest speed, about 3s/it or 1:17 per second of video w/ a 4090. Best quality will be obtained with just block swapping/sdpa attention/full model though.
|
62 |
-
|
63 |
-
Put loras in a /lora subfolder, if not trained with musubi you need to convert them.
|
64 |
-
|
65 |
-
Only unipc is supported for now. Sage attn is experimental. When using the F1 model not all options available for the original framepack model will work, like endframe and sectional images.
|
66 |
-
|
67 |
-
Here is an example prompt for a 5 second video with 4 sections using sectional prompting, also supports longer videos with indexes ie 0-2 ;;;3-5 etc:
|
68 |
-
|
69 |
-
0:A cinematic video showcases a cute blue penguin wearing sunglasses. The penguin runs quickly into mcdonalds.;;;1:The penguin runs quickly into mcdonalds and jumps up on a table and starts eating his food. The penguin's name is Piplup he is a famous Pokemon actor. The video is a fast action sequence animation showing the penguin running into a mcdonalds an jumping up onto a table.;;;2:The penguin is seated at a table and is enjoying his happy meal. The penguin's name is Piplup he is a famous Pokemon actor. The video is a fast action sequence animation showing the penguin running into a mcdonalds and jumping up onto a table.;;;3:The penguin is seated at a table and is happily enjoying his happy meal. The penguin's name is Piplup he is a famous Pokemon actor. The penguin flexes his huge arm muscles at the end of the video.
|
70 |
-
|
71 |
-
I have added support for 4 sectional images during inference. It works best when the images are close together. Refer to the screen shot for an example of a working 5 second video.
|
72 |
-
|
73 |
-
For more details on using framepack with musubi go here https://github.com/kohya-ss/musubi-tuner/blob/main/docs/framepack.md
|
74 |
-
|
75 |
-
Fastest speed will be achieved with fp8 and fp8 scaled, then you can reduce block swapping to your memory constraints. (leave about 1gb free)
|
76 |
-
|
77 |
-
Framepack Extension tab is still a work in progress.
|
78 |
-
Thanks to @pftq https://github.com/pftq and @chaojie https://github.com/chaojie for their work on the extension logics.
|
79 |
-
|
80 |
-
## To Use the new Skyreels-V2 models
|
81 |
-
|
82 |
-
I have provided these 2 at https://huggingface.co/maybleMyers/wan_files_for_h1111
|
83 |
-
|
84 |
-
SkyReels-V2-I2V-14B-720P-FP16.safetensors
|
85 |
-
SkyReels-V2-I2V-14B-540P-FP16.safetensors
|
86 |
-
|
87 |
-
You can just drop them into the wan folder and use them in the WanX-i2v tab. Skyreels-V2 is a fine tune from Wan2.1.
|
88 |
-
If you have download the kijai variants the will not work because he added extra keys to the model.
|
89 |
-
|
90 |
-
## To Use WanX
|
91 |
-
|
92 |
-
To use wanX download these and toss them in the wan subfolder:
|
93 |
-
Download the T5 `models_t5_umt5-xxl-enc-bf16.pth`, vae `Wan2.1_VAE.pth` and CLIP `models_clip_open-clip-xlm-roberta-large-vit-huge-14.pth` from the following page: https://huggingface.co/Wan-AI/Wan2.1-I2V-14B-720P/tree/main
|
94 |
-
|
95 |
-
Download the DiT weights from the following page: https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/tree/main/split_files/diffusion_models
|
96 |
-
ie : wan2.1_i2v_720p_14B_fp16.safetensors
|
97 |
-
|
98 |
-
For the fun control option in WanX-i2v I recommend the fp16 weights here: https://huggingface.co/maybleMyers/wan_files_for_h1111/tree/main
|
99 |
-
Wan2.1-Fun-14B-Control_fp16.safetensors
|
100 |
-
|
101 |
-
git pull to update the installation
|
102 |
-
pip install -r requirements.txt
|
103 |
-
|
104 |
-
I have tested the 14B i2v and t2v models so far to be working
|
105 |
-
|
106 |
-
## Requirements
|
107 |
-
|
108 |
-
- Python 3.10
|
109 |
-
- CUDA 12.4
|
110 |
-
|
111 |
-
## Basic Installation (Linux)
|
112 |
-
|
113 |
-
Tested on ubuntu 24
|
114 |
-
|
115 |
-
to update navigate to H1111 and git pull
|
116 |
-
|
117 |
-
```powershell
|
118 |
-
git clone https://github.com/maybleMyers/H1111
|
119 |
-
cd H1111
|
120 |
-
python -m venv env
|
121 |
-
#(if you have another version of python do python3.10 -m venv env after you install it with sudo apt install python3.10 python3.10-venv python3.10-distutils)
|
122 |
-
source env/bin/activate
|
123 |
-
pip install torch==2.5.1 torchvision --index-url https://download.pytorch.org/whl/cu124
|
124 |
-
pip install -r requirements.txt
|
125 |
-
pip install flash-attn --no-build-isolation
|
126 |
-
pip install sageattention==1.0.6
|
127 |
-
might need python3.10-dev as well for sage attention to work
|
128 |
-
|
129 |
-
```
|
130 |
-
|
131 |
-
run with
|
132 |
-
source env/bin/activate
|
133 |
-
python h1111.py
|
134 |
-
|
135 |
-
for GPU1
|
136 |
-
CUDA_VISIBLE_DEVICES=1 python h1111.py
|
137 |
-
|
138 |
-
## Basic Installation (Windows)
|
139 |
-
|
140 |
-
|
141 |
-
|
142 |
-
First, open PowerShell and navigate to your desired installation directory. Then run these commands:
|
143 |
-
|
144 |
-
```powershell
|
145 |
-
git clone https://github.com/maybleMyers/H1111
|
146 |
-
cd H1111
|
147 |
-
python -m venv env
|
148 |
-
./env/scripts/activate
|
149 |
-
pip install torch==2.5.1 torchvision --index-url https://download.pytorch.org/whl/cu124
|
150 |
-
pip install -r requirements.txt
|
151 |
-
|
152 |
-
```
|
153 |
-
|
154 |
-
## To run
|
155 |
-
|
156 |
-
```
|
157 |
-
env/scripts/activate
|
158 |
-
python h1111.py
|
159 |
-
```
|
160 |
-
|
161 |
-
open 127.0.0.1:7860 in a browser
|
162 |
-
|
163 |
-
You can set cuda device to 1,2,3,4,5,6,7 etc in the env once activated in a separate terminal to run unlimited copies at once if you have another gpu.
|
164 |
-
ie for linux on the second gpu: CUDA_VISIBLE_DEVICES=1 python h1111.py
|
165 |
-
|
166 |
-
## full changlog
|
167 |
-
5/25/2025
|
168 |
-
Enable full intermediate previews for framepack tab, some change to framepack extension with image input logic.
|
169 |
-
5/24/2025
|
170 |
-
Batch images from folder now available for framepack. Save only extension option and experimental start image for f1 in framepack extension tab.
|
171 |
-
5/23/2025
|
172 |
-
Added ability to use the normal framepack model with endframe support in the framepack extension tab. Support additional bucket sizes.
|
173 |
-
5/18/2025
|
174 |
-
Add video extension tab. Currently only works with f1 model. Full credit to @pfqt and @chaojie for their amazing work!
|
175 |
-
5/12/2025
|
176 |
-
Add skip button to framepack.
|
177 |
-
5/9/2025
|
178 |
-
Add testing branch for framepack F1 end image, kinda glitchygo https://github.com/maybleMyers/H1111/tree/f1_end
|
179 |
-
5/5/2025
|
180 |
-
Update an experimental hunyuan to framepack convert lora option in the convert lora tab.
|
181 |
-
Add tea cache to frame pack.
|
182 |
-
5/3/2025
|
183 |
-
Add support for framepack F1! download from https://huggingface.co/maybleMyers/wan_files_for_h1111/blob/main/FramePack_F1_I2V_HY_20250503.safetensors put it in your hunyuan folder. You might need to reinstall reqs "pip install -r requirements.txt"
|
184 |
-
Add support for Wan2.1 i2v-14B-FC-1.1. It is a fun control model and is very good. Use it in the WanX-i2v tab and make sure to select the task i2v-14B-FC-1.1 at the bottom of the page. Download the weights from https://huggingface.co/maybleMyers/wan_files_for_h1111
|
185 |
-
4/30/2025
|
186 |
-
Previews for framepack.
|
187 |
-
4/29/2025
|
188 |
-
Add initial preview support to the wanX-i2v tab based. If you want to use them use the preview branch. Thanks to Sarania.
|
189 |
-
Wan2.1-Fun-V1.1-14B-InP-FP16.safetensors is available at https://huggingface.co/maybleMyers/wan_files_for_h1111
|
190 |
-
Fix bug in hunyuan-t2v not loading lora.
|
191 |
-
4/26/2025
|
192 |
-
Add SkyReels-V2-I2V-14B-720P-FP16.safetensors to supported models.
|
193 |
-
Added alot better options for Framepack including working sectional images, Thanks to kohya!
|
194 |
-
4/25/2025
|
195 |
-
Framepack backend updates for better LoRa support for LoRa's trained with musubi tuner. Also better weighting options.
|
196 |
-
4/24/2025
|
197 |
-
Update FramePack backend to musubi backend instead of original. Offers much improved speed and some quality improvements.
|
198 |
-
Add support for torch 2.7.0 + cuda 12.8
|
199 |
-
4/18/2025
|
200 |
-
Add initial support for FramePack. https://github.com/lllyasviel/FramePack
|
201 |
-
4/15/2025
|
202 |
-
Add much improved functionality for the wan fun control model. Added strength imrpovements and dropoff code to choose when to apply the control video. Thanks wordbrew.
|
203 |
-
4/3/2025
|
204 |
-
Add support for hunyuan i2v model. Download the clip vision from https://huggingface.co/maybleMyers/H1111_Hunyuan_i2v And download the official model from hunyuan's website and rename it to mp_rank_00_model_states_i2v.pt https://huggingface.co/tencent/HunyuanVideo-I2V/tree/main/hunyuan-video-i2v-720p/transformers add both to your hunyuan folder.
|
205 |
-
3/29/2025
|
206 |
-
Added support for fun models! download dit from https://huggingface.co/alibaba-pai/Wan2.1-Fun-14B-Control and specify correct task type and dit location. I renamed it from diffusion_pytorch_model to Wan2.1-Fun-14B-control. Works in the normal WanX-i2v tab when you select the control option at the bottom of the page.
|
207 |
-
3/23/2025
|
208 |
-
Added Wanx cfg skip functionality to skip cfg guidance during inference for faster generations but less following of the prompt
|
209 |
-
3/22/2025
|
210 |
-
Added WanX-i2v end frame functionality
|
211 |
-
3/20/2025
|
212 |
-
Added WanX-v2v functionality.
|
213 |
-
3/18/2025
|
214 |
-
Added Skip Layer Guidance for WanX-i2v.
|
215 |
-
3/13/2025
|
216 |
-
Added extend video functionality to WanX-i2v. It kind of works .
|
217 |
-
3/12/2025
|
218 |
-
Added ability to send the last frame of a video to the input in WanX-i2v. Also you can now use this to extend the video. You can do multiple batches at each step and pick the best extended video then generate an even longer one.
|
219 |
-
3/9/2025
|
220 |
-
Added batching ability for a folder full of images in WanX-i2v tab. Added flash attn for windows prebuilt wheel.
|
221 |
-
3/8/2025
|
222 |
-
Added support for wan lora's. Remember to convert them first in the convert lora tab.
|
223 |
-
3/5/2025
|
224 |
-
Added ability to batch a folder of images with skyreels i2v, so you can make a video with every image in a folder.
|
225 |
-
3/2/2025
|
226 |
-
Added initial support for wanX-2.1 Image to Video and Text to Video inference.
|
227 |
-
3/1/2025
|
228 |
-
Added support for Skyreels Video to Video and Text to Video.
|
229 |
-
2/23/2025
|
230 |
-
Added initial support for skyreels-V1 using musubi's skyreel implementation. (thanks sdbds)
|
231 |
-
download models from https://huggingface.co/Kijai/SkyReels-V1-Hunyuan_comfy and add them to your hunyuan folder
|
232 |
-
skyreels_hunyuan_i2v_bf16.safetensors
|
233 |
-
skyreels_hunyuan_t2v_bf16.safetensors
|
234 |
-
|
235 |
-
|
236 |
-
## to use stock hunyuan models
|
237 |
-
|
238 |
-
https://huggingface.co/tencent/HunyuanVideo/resolve/main/hunyuan-video-t2v-720p/transformers/mp_rank_00_model_states.pt
|
239 |
-
|
240 |
-
https://huggingface.co/tencent/HunyuanVideo/resolve/main/hunyuan-video-t2v-720p/vae/pytorch_model.pt
|
241 |
-
|
242 |
-
https://huggingface.co/Comfy-Org/HunyuanVideo_repackaged/resolve/main/split_files/text_encoders/llava_llama3_fp16.safetensors
|
243 |
-
|
244 |
-
https://huggingface.co/Comfy-Org/HunyuanVideo_repackaged/resolve/main/split_files/text_encoders/clip_l.safetensors
|
245 |
-
|
246 |
-
#fp8 dit model
|
247 |
-
|
248 |
-
https://huggingface.co/kohya-ss/HunyuanVideo-fp8_e4m3fn-unofficial/resolve/main/mp_rank_00_model_states_fp8.safetensors
|
249 |
-
|
250 |
-
place models in H1111/hunyuan folder
|
251 |
-
|
252 |
-
### Optional: Install Xformers
|
253 |
-
```powershell
|
254 |
-
pip install --no-deps xformers --index-url https://download.pytorch.org/whl/cu124
|
255 |
-
```
|
256 |
-
|
257 |
-
### Optional: Install Flash Attention
|
258 |
-
Note: This can take 1-5 hour to install even on a good CPU, but provides faster generation.
|
259 |
-
I have uploaded a wheel for windows users to match cuda 12.4 and python 3.10.(thanks lldacing)
|
260 |
-
https://huggingface.co/maybleMyers/wan_files_for_h1111/resolve/main/flash_attn-2.7.4%2Bcu124torch2.5.1cxx11abiFALSE-cp310-cp310-win_amd64.whl?download=true
|
261 |
-
|
262 |
-
```powershell
|
263 |
-
pip install flash-attn --no-build-isolation
|
264 |
-
|
265 |
-
If you have downloaded the wheel you can install it with:
|
266 |
-
|
267 |
-
pip install "flash_attn-2.7.4+cu124torch2.5.1cxx11abiFALSE-cp310-cp310-win_amd64.whl"
|
268 |
-
```
|
269 |
-
```
|
|
|
1 |
+
---
|
2 |
+
sdk: gradio
|
3 |
+
title: Framepack STUDIO H111
|
4 |
+
app_file: studio3.py
|
5 |
+
pinned: false
|
6 |
+
license: apache-2.0
|
7 |
+
short_description: fast video generation from images & text
|
8 |
+
sdk_version: 5.31.0
|
9 |
+
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|