rahul7star commited on
Commit
0487312
·
verified ·
1 Parent(s): e0336bc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -269
README.md CHANGED
@@ -1,269 +1,9 @@
1
- ![GUI Screenshot](images/screenshot.png)
2
-
3
- # Recent update
4
- 5/25/2025
5
- Enable full intermediate previews for framepack tab, some change to framepack extension with image input logic.
6
- 5/24/2025
7
- Batch images from folder now available for framepack. Save only extension option and experimental start image for f1 in framepack extension tab.
8
- 5/23/2025
9
- Added ability to use the normal framepack model with endframe support in the framepack extension tab. Support additional bucket sizes.
10
- 5/18/2025
11
- Add video extension tab. Currently only works with f1 model. Full credit to @pfqt and @chaojie for their amazing work!
12
-
13
- # H1111
14
-
15
- This is a GUI for tech wizard kohya-ss's musubi tuner's inference script.
16
- https://github.com/kohya-ss/musubi-tuner
17
-
18
- It allows inference with these models:
19
- FramePack
20
- Hunyuan-t2v
21
- Hunyuan-i2v
22
- Hunyuan-v2v
23
- WanX-t2v
24
- WanX-i2v
25
- WanX-v2v
26
- SkyReels-i2v
27
- SkyReels-t2v
28
-
29
- I have mostly been workiing on the framepack tab and the WanX-i2v tab. They are the best to use right now. WanX-i2v is used for skyreels v2 and the fun control models.
30
-
31
- This supports queuing multiple different jobs if you open 2+ browser tabs and use the same model.
32
-
33
- If you are running out of vram use more block swapping. Using FP8 scaled is also a decent option to lower memory usage, select fp8 and fp8 scaled to use it. Scaled fp8 tries to duplicate the important parts of the model from FP16. Sage attention is the fastest/lowest vram but difficult to install in windows.
34
-
35
- Best quality will be obtained with only enabling block swapping and using the fp16 model with sdpa attention. You can speed things up with cfg skip, fp8 scaled, slg skip is small speedup, sage attention is fastest but all speedups come with quality degradations. I designed this to try to focus on quality over speed.
36
-
37
- If you are using a lora that you didn't train with musubi you need to drag it to the convert lora tab and convert it to the default format. It should spit it out into the /lora folder.
38
-
39
- If you need additional installation instructions or information create an issue and I will try to help. Also there are alot of settings notes on the musubi github linked above.
40
-
41
- For torch 2.7.0 and windows installation try:
42
- pip install typing-extensions
43
- pip install torch==2.7.0+cu128 torchvision==0.22.0+cu128 --index-url https://download.pytorch.org/whl/cu128
44
- pip install -r requirementsTorch27.txt
45
-
46
- ## To Use FramePack
47
-
48
-
49
- download these 5 files from https://huggingface.co/maybleMyers/framepack_h1111 and put them in a subfolder named hunyuan (H1111/hunyuan), or reference where they are in the gui if you have already aquired them.
50
-
51
- FramePackI2V_HY_bf16.safetensors or FramePack_F1_I2V_HY_20250503.safetensors for F1
52
-
53
- clip_l.safetensors
54
-
55
- llava_llama3_fp16.safetensors
56
-
57
- model.safetensors
58
-
59
- pytorch_model.pt
60
-
61
- Lora trained with musubi tuner's framepack training confirmed to work great. Normal lora trained for hunyuan kinda suck. Use a lot of block swap this is a different back end than the official repo. If you select fp8 and fp8 scaled it will all fit on a 24gb gpu for fastest speed, about 3s/it or 1:17 per second of video w/ a 4090. Best quality will be obtained with just block swapping/sdpa attention/full model though.
62
-
63
- Put loras in a /lora subfolder, if not trained with musubi you need to convert them.
64
-
65
- Only unipc is supported for now. Sage attn is experimental. When using the F1 model not all options available for the original framepack model will work, like endframe and sectional images.
66
-
67
- Here is an example prompt for a 5 second video with 4 sections using sectional prompting, also supports longer videos with indexes ie 0-2 ;;;3-5 etc:
68
-
69
- 0:A cinematic video showcases a cute blue penguin wearing sunglasses. The penguin runs quickly into mcdonalds.;;;1:The penguin runs quickly into mcdonalds and jumps up on a table and starts eating his food. The penguin's name is Piplup he is a famous Pokemon actor. The video is a fast action sequence animation showing the penguin running into a mcdonalds an jumping up onto a table.;;;2:The penguin is seated at a table and is enjoying his happy meal. The penguin's name is Piplup he is a famous Pokemon actor. The video is a fast action sequence animation showing the penguin running into a mcdonalds and jumping up onto a table.;;;3:The penguin is seated at a table and is happily enjoying his happy meal. The penguin's name is Piplup he is a famous Pokemon actor. The penguin flexes his huge arm muscles at the end of the video.
70
-
71
- I have added support for 4 sectional images during inference. It works best when the images are close together. Refer to the screen shot for an example of a working 5 second video.
72
-
73
- For more details on using framepack with musubi go here https://github.com/kohya-ss/musubi-tuner/blob/main/docs/framepack.md
74
-
75
- Fastest speed will be achieved with fp8 and fp8 scaled, then you can reduce block swapping to your memory constraints. (leave about 1gb free)
76
-
77
- Framepack Extension tab is still a work in progress.
78
- Thanks to @pftq https://github.com/pftq and @chaojie https://github.com/chaojie for their work on the extension logics.
79
-
80
- ## To Use the new Skyreels-V2 models
81
-
82
- I have provided these 2 at https://huggingface.co/maybleMyers/wan_files_for_h1111
83
-
84
- SkyReels-V2-I2V-14B-720P-FP16.safetensors
85
- SkyReels-V2-I2V-14B-540P-FP16.safetensors
86
-
87
- You can just drop them into the wan folder and use them in the WanX-i2v tab. Skyreels-V2 is a fine tune from Wan2.1.
88
- If you have download the kijai variants the will not work because he added extra keys to the model.
89
-
90
- ## To Use WanX
91
-
92
- To use wanX download these and toss them in the wan subfolder:
93
- Download the T5 `models_t5_umt5-xxl-enc-bf16.pth`, vae `Wan2.1_VAE.pth` and CLIP `models_clip_open-clip-xlm-roberta-large-vit-huge-14.pth` from the following page: https://huggingface.co/Wan-AI/Wan2.1-I2V-14B-720P/tree/main
94
-
95
- Download the DiT weights from the following page: https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/tree/main/split_files/diffusion_models
96
- ie : wan2.1_i2v_720p_14B_fp16.safetensors
97
-
98
- For the fun control option in WanX-i2v I recommend the fp16 weights here: https://huggingface.co/maybleMyers/wan_files_for_h1111/tree/main
99
- Wan2.1-Fun-14B-Control_fp16.safetensors
100
-
101
- git pull to update the installation
102
- pip install -r requirements.txt
103
-
104
- I have tested the 14B i2v and t2v models so far to be working
105
-
106
- ## Requirements
107
-
108
- - Python 3.10
109
- - CUDA 12.4
110
-
111
- ## Basic Installation (Linux)
112
-
113
- Tested on ubuntu 24
114
-
115
- to update navigate to H1111 and git pull
116
-
117
- ```powershell
118
- git clone https://github.com/maybleMyers/H1111
119
- cd H1111
120
- python -m venv env
121
- #(if you have another version of python do python3.10 -m venv env after you install it with sudo apt install python3.10 python3.10-venv python3.10-distutils)
122
- source env/bin/activate
123
- pip install torch==2.5.1 torchvision --index-url https://download.pytorch.org/whl/cu124
124
- pip install -r requirements.txt
125
- pip install flash-attn --no-build-isolation
126
- pip install sageattention==1.0.6
127
- might need python3.10-dev as well for sage attention to work
128
-
129
- ```
130
-
131
- run with
132
- source env/bin/activate
133
- python h1111.py
134
-
135
- for GPU1
136
- CUDA_VISIBLE_DEVICES=1 python h1111.py
137
-
138
- ## Basic Installation (Windows)
139
-
140
-
141
-
142
- First, open PowerShell and navigate to your desired installation directory. Then run these commands:
143
-
144
- ```powershell
145
- git clone https://github.com/maybleMyers/H1111
146
- cd H1111
147
- python -m venv env
148
- ./env/scripts/activate
149
- pip install torch==2.5.1 torchvision --index-url https://download.pytorch.org/whl/cu124
150
- pip install -r requirements.txt
151
-
152
- ```
153
-
154
- ## To run
155
-
156
- ```
157
- env/scripts/activate
158
- python h1111.py
159
- ```
160
-
161
- open 127.0.0.1:7860 in a browser
162
-
163
- You can set cuda device to 1,2,3,4,5,6,7 etc in the env once activated in a separate terminal to run unlimited copies at once if you have another gpu.
164
- ie for linux on the second gpu: CUDA_VISIBLE_DEVICES=1 python h1111.py
165
-
166
- ## full changlog
167
- 5/25/2025
168
- Enable full intermediate previews for framepack tab, some change to framepack extension with image input logic.
169
- 5/24/2025
170
- Batch images from folder now available for framepack. Save only extension option and experimental start image for f1 in framepack extension tab.
171
- 5/23/2025
172
- Added ability to use the normal framepack model with endframe support in the framepack extension tab. Support additional bucket sizes.
173
- 5/18/2025
174
- Add video extension tab. Currently only works with f1 model. Full credit to @pfqt and @chaojie for their amazing work!
175
- 5/12/2025
176
- Add skip button to framepack.
177
- 5/9/2025
178
- Add testing branch for framepack F1 end image, kinda glitchygo https://github.com/maybleMyers/H1111/tree/f1_end
179
- 5/5/2025
180
- Update an experimental hunyuan to framepack convert lora option in the convert lora tab.
181
- Add tea cache to frame pack.
182
- 5/3/2025
183
- Add support for framepack F1! download from https://huggingface.co/maybleMyers/wan_files_for_h1111/blob/main/FramePack_F1_I2V_HY_20250503.safetensors put it in your hunyuan folder. You might need to reinstall reqs "pip install -r requirements.txt"
184
- Add support for Wan2.1 i2v-14B-FC-1.1. It is a fun control model and is very good. Use it in the WanX-i2v tab and make sure to select the task i2v-14B-FC-1.1 at the bottom of the page. Download the weights from https://huggingface.co/maybleMyers/wan_files_for_h1111
185
- 4/30/2025
186
- Previews for framepack.
187
- 4/29/2025
188
- Add initial preview support to the wanX-i2v tab based. If you want to use them use the preview branch. Thanks to Sarania.
189
- Wan2.1-Fun-V1.1-14B-InP-FP16.safetensors is available at https://huggingface.co/maybleMyers/wan_files_for_h1111
190
- Fix bug in hunyuan-t2v not loading lora.
191
- 4/26/2025
192
- Add SkyReels-V2-I2V-14B-720P-FP16.safetensors to supported models.
193
- Added alot better options for Framepack including working sectional images, Thanks to kohya!
194
- 4/25/2025
195
- Framepack backend updates for better LoRa support for LoRa's trained with musubi tuner. Also better weighting options.
196
- 4/24/2025
197
- Update FramePack backend to musubi backend instead of original. Offers much improved speed and some quality improvements.
198
- Add support for torch 2.7.0 + cuda 12.8
199
- 4/18/2025
200
- Add initial support for FramePack. https://github.com/lllyasviel/FramePack
201
- 4/15/2025
202
- Add much improved functionality for the wan fun control model. Added strength imrpovements and dropoff code to choose when to apply the control video. Thanks wordbrew.
203
- 4/3/2025
204
- Add support for hunyuan i2v model. Download the clip vision from https://huggingface.co/maybleMyers/H1111_Hunyuan_i2v And download the official model from hunyuan's website and rename it to mp_rank_00_model_states_i2v.pt https://huggingface.co/tencent/HunyuanVideo-I2V/tree/main/hunyuan-video-i2v-720p/transformers add both to your hunyuan folder.
205
- 3/29/2025
206
- Added support for fun models! download dit from https://huggingface.co/alibaba-pai/Wan2.1-Fun-14B-Control and specify correct task type and dit location. I renamed it from diffusion_pytorch_model to Wan2.1-Fun-14B-control. Works in the normal WanX-i2v tab when you select the control option at the bottom of the page.
207
- 3/23/2025
208
- Added Wanx cfg skip functionality to skip cfg guidance during inference for faster generations but less following of the prompt
209
- 3/22/2025
210
- Added WanX-i2v end frame functionality
211
- 3/20/2025
212
- Added WanX-v2v functionality.
213
- 3/18/2025
214
- Added Skip Layer Guidance for WanX-i2v.
215
- 3/13/2025
216
- Added extend video functionality to WanX-i2v. It kind of works .
217
- 3/12/2025
218
- Added ability to send the last frame of a video to the input in WanX-i2v. Also you can now use this to extend the video. You can do multiple batches at each step and pick the best extended video then generate an even longer one.
219
- 3/9/2025
220
- Added batching ability for a folder full of images in WanX-i2v tab. Added flash attn for windows prebuilt wheel.
221
- 3/8/2025
222
- Added support for wan lora's. Remember to convert them first in the convert lora tab.
223
- 3/5/2025
224
- Added ability to batch a folder of images with skyreels i2v, so you can make a video with every image in a folder.
225
- 3/2/2025
226
- Added initial support for wanX-2.1 Image to Video and Text to Video inference.
227
- 3/1/2025
228
- Added support for Skyreels Video to Video and Text to Video.
229
- 2/23/2025
230
- Added initial support for skyreels-V1 using musubi's skyreel implementation. (thanks sdbds)
231
- download models from https://huggingface.co/Kijai/SkyReels-V1-Hunyuan_comfy and add them to your hunyuan folder
232
- skyreels_hunyuan_i2v_bf16.safetensors
233
- skyreels_hunyuan_t2v_bf16.safetensors
234
-
235
-
236
- ## to use stock hunyuan models
237
-
238
- https://huggingface.co/tencent/HunyuanVideo/resolve/main/hunyuan-video-t2v-720p/transformers/mp_rank_00_model_states.pt
239
-
240
- https://huggingface.co/tencent/HunyuanVideo/resolve/main/hunyuan-video-t2v-720p/vae/pytorch_model.pt
241
-
242
- https://huggingface.co/Comfy-Org/HunyuanVideo_repackaged/resolve/main/split_files/text_encoders/llava_llama3_fp16.safetensors
243
-
244
- https://huggingface.co/Comfy-Org/HunyuanVideo_repackaged/resolve/main/split_files/text_encoders/clip_l.safetensors
245
-
246
- #fp8 dit model
247
-
248
- https://huggingface.co/kohya-ss/HunyuanVideo-fp8_e4m3fn-unofficial/resolve/main/mp_rank_00_model_states_fp8.safetensors
249
-
250
- place models in H1111/hunyuan folder
251
-
252
- ### Optional: Install Xformers
253
- ```powershell
254
- pip install --no-deps xformers --index-url https://download.pytorch.org/whl/cu124
255
- ```
256
-
257
- ### Optional: Install Flash Attention
258
- Note: This can take 1-5 hour to install even on a good CPU, but provides faster generation.
259
- I have uploaded a wheel for windows users to match cuda 12.4 and python 3.10.(thanks lldacing)
260
- https://huggingface.co/maybleMyers/wan_files_for_h1111/resolve/main/flash_attn-2.7.4%2Bcu124torch2.5.1cxx11abiFALSE-cp310-cp310-win_amd64.whl?download=true
261
-
262
- ```powershell
263
- pip install flash-attn --no-build-isolation
264
-
265
- If you have downloaded the wheel you can install it with:
266
-
267
- pip install "flash_attn-2.7.4+cu124torch2.5.1cxx11abiFALSE-cp310-cp310-win_amd64.whl"
268
- ```
269
- ```
 
1
+ ---
2
+ sdk: gradio
3
+ title: Framepack STUDIO H111
4
+ app_file: studio3.py
5
+ pinned: false
6
+ license: apache-2.0
7
+ short_description: fast video generation from images & text
8
+ sdk_version: 5.31.0
9
+ ---