GGUF Quants of Wan 2.1 for lower VRAM usage.
I suggest to use TeaCache for ~2x speedup.
All samples are 640x640 with TeaCache enabled and a Init-Image.
Source: Source (gguf): https://huggingface.co/city96/Wan2.1-I2V-14B-480P-gguf/tree/main from city96
Here are the results of some testing ...
VRAM and RAM calculations:
Your VRAM should be 1~2 GB higher as the model size
e.g. 16GB (VRAM): Model 12GB -2GB = 10 -> good to go
You should always spare 1-2 GB for you OS to operate functional
Your RAM should at least be 16, better 32 GB
Hints for faster but reasonable good results I2V (Image to Video):
Use a low resolution inside the maximum pixel count
Upscale the complete video after generation with a tool of your liking
Use 20-30 steps, 20 is good, 30 needs significant more time but preserves details
The step increase is linear so, 10 more steps results from e.g. 10 min to 15 min
Use CGF 4-6 (6 are good most of the time), 4 are sometimes better if lighting goes off
The image as Init-Image you use should have the same ratio as the video resolution
Good resolutions:
9:16/ 16:9 with 480x832 / 832x480 = 480p
3:4 / 4:3 with 480x640 / 640x480 = 480p
Wan-AI / Wan2.1 Video Model (Safetensors) - GGUF Quants - i2v 14B is a highly specialized Image generation AI Model of type Safetensors / Checkpoint AI Model created by AI community user darksidewalker. Derived from the powerful Stable Diffusion (Wan Video) model, Wan-AI / Wan2.1 Video Model (Safetensors) - GGUF Quants - i2v 14B has undergone an extensive fine-tuning process, leveraging the power of a dataset consisting of images generated by other AI models or user-contributed data. This fine-tuning process ensures that Wan-AI / Wan2.1 Video Model (Safetensors) - GGUF Quants - i2v 14B is capable of generating images that are highly relevant to the specific use-cases it was designed for, such as base model.
With a rating of 0 and over 0 ratings, Wan-AI / Wan2.1 Video Model (Safetensors) - GGUF Quants - i2v 14B is a popular choice among users for generating high-quality images from text prompts.
Yes! You can download the latest version of Wan-AI / Wan2.1 Video Model (Safetensors) - GGUF Quants - i2v 14B from here.
To use Wan-AI / Wan2.1 Video Model (Safetensors) - GGUF Quants - i2v 14B, download the model checkpoint file and set up an UI for running Stable Diffusion models (for example, AUTOMATIC1111). Then, provide the model with a detailed text prompt to generate an image. Experiment with different prompts and settings to achieve the desired results. If this sounds a bit complicated, check out our initial guide to Stable Diffusion – it might be of help. And if you really want to dive deep into AI image generation and understand how set up AUTOMATIC1111 to use Safetensors / Checkpoint AI Models like Wan-AI / Wan2.1 Video Model (Safetensors) - GGUF Quants - i2v 14B, check out our crash course in AI image generation.
Image2Video for under 16GB VRAM
Go ahead and upload yours!
Your query returned no results – please try removing some filters or trying a different term.