Note: There are other Wan Video files hosted on Civitai - these may be duplicates, but this model card is primarily to host the files used by Wan Video in the Civitai Generator.
Wan2.2, a major upgrade to our visual generative models, which is now open-sourced, offering more powerful capabilities, better performance, and superior visual quality. With Wan2.2, we have focused on incorporating the following technical innovations:
π MoE Architecture: Wan2.2 introduces a Mixture-of-Experts (MoE) architecture into video diffusion models. By separating the denoising process cross timesteps with specialized powerful expert models, this enlarges the overall model capacity while maintaining the same computational cost.
πͺπ» Data Scaling: Compared to Wan2.1, Wan2.2 is trained on a significantly larger data, with +65.6% more images and +83.2% more videos. This expansion notably enhances the model's generalization across multiple dimensions such as motions, semantics, and aesthetics, achieving TOP performance among all open-sourced and closed-sourced models.
π¬ Cinematic Aesthetics: Wan2.2 incorporates specially curated aesthetic data with fine-grained labels for lighting, composition, and color. This allows for more precise and controllable cinematic style generation, facilitating the creation of videos with customizable aesthetic preferences.
π Efficient High-Definition Hybrid TI2V: Wan2.2 open-sources a 5B model built with our advanced Wan2.2-VAE that achieves a compression ratio of 16Γ16Γ4. This model supports both text-to-video and image-to-video generation at 720P resolution with 24fps and can also run on consumer-grade graphics cards like 4090. It one of the fastest 720P@24fps models currently available, capable of serving both the industrial and academic sectors simultaneously.
The T2V-A14B model, supports generating 5s videos at both 480P and 720P resolutions. Built with a Mixture-of-Experts (MoE) architecture, it delivers outstanding video generation quality. On our new benchmark Wan-Bench 2.0, the model surpasses leading commercial models across most key evaluation dimensions.
The I2V-A14B model, designed for image-to-video generation, supports both 480P and 720P resolutions. Built with a Mixture-of-Experts (MoE) architecture, it achieves more stable video synthesis with reduced unrealistic camera movements and offers enhanced support for diverse stylized scenes.
The TI2V-5B model is built with the advanced Wan2.2-VAE that achieves a compression ratio of 16Γ16Γ4. This model supports both text-to-video and image-to-video generation at 720P resolution with 24fps and can runs on single consumer-grade GPU such as the 4090. It is one of the fastest 720P@24fps models available, meeting the needs of both industrial applications and academic research.
GitHub: https://github.com/Wan-Video/Wan2.2
Originally HuggingFace Repo: https://huggingface.co/Comfy-Org/Wan_2.2_ComfyUI_Repackaged/tree/main/split_files/diffusion_models
Wan Video 2.2 - ComfyUI Repack is a highly specialized Image generation AI Model of type Safetensors / Checkpoint AI Model created by AI community user theally. Derived from the powerful Stable Diffusion (Wan Video 2.2 TI2V-5B) model, Wan Video 2.2 - ComfyUI Repack has undergone an extensive fine-tuning process, leveraging the power of a dataset consisting of images generated by other AI models or user-contributed data. This fine-tuning process ensures that Wan Video 2.2 - ComfyUI Repack is capable of generating images that are highly relevant to the specific use-cases it was designed for, such as base model.
With a rating of 0 and over 0 ratings, Wan Video 2.2 - ComfyUI Repack is a popular choice among users for generating high-quality images from text prompts.
Yes! You can download the latest version of Wan Video 2.2 - ComfyUI Repack from here.
To use Wan Video 2.2 - ComfyUI Repack, download the model checkpoint file and set up an UI for running Stable Diffusion models (for example, AUTOMATIC1111). Then, provide the model with a detailed text prompt to generate an image. Experiment with different prompts and settings to achieve the desired results. If this sounds a bit complicated, check out our initial guide to Stable Diffusion β it might be of help. And if you really want to dive deep into AI image generation and understand how set up AUTOMATIC1111 to use Safetensors / Checkpoint AI Models like Wan Video 2.2 - ComfyUI Repack, check out our crash course in AI image generation.
Wan 2.2 14B for Text-to-Video on-site Generation
Go ahead and upload yours!
Your query returned no results β please try removing some filters or trying a different term.