This workflow is my take on organizing and building an I2V model running against a "smaller" graphics card (specifically 12 GB). I use Q5 GGUF models, though I have also tested against smaller Q4 and Q3 - each having a slight decrease to prompt adherence and quality, but still usable. This includes using the lightspeed lora and notes on where all models / loras can be downloaded from.
I suggest if you have limited RAM or VRAM, to run ComfyUI with the parameter: --cache-none
While this will mean multiple batches against the same video will be slower, you get a much more consistent overall generation speed of your videos (3-4 minutes for 5-6 seconds on moderate home PC configurations).
This also includes using Florence2 (LLM) for image detection and auto-prompt assistance. You only need to add your action to the manual prompt (if desired).
There are a lot of nodes I have seen in various workflows ... but in Wan 2.2 I2V, at least, they tend to have no effect and only increase overhead.
I run my videos typically in 480p @ 480 x 832, and this workflow then upscales by 2x to 960 x 1664.
ComfyUI-GGUF (https://github.com/city96/ComfyUI-GGUF)
rgthree-comfy (https://github.com/rgthree/rgthree-comfy)
ComfyUI-KJNodes (https://github.com/kijai/ComfyUI-KJNodes)
ComfyUI-Florence2 (https://github.com/kijai/ComfyUI-Florence2)
ComfyUI-VideoHelperSuite (https://github.com/Kosinkadink/ComfyUI-VideoHelperSuite)
WAS Node Suite (https://github.com/WASasquatch/was-node-suite-comfyui)
Minor Updates
Updated nodes, refreshed (including Wan updates for improved coloring)
Removed obscure node for file naming
Go ahead and upload yours!
Your query returned no results β please try removing some filters or trying a different term.