ComfyUITemplates.com

Discover free ready-made ComfyUI templates for AI workflows.

Wan Animate Pose transfer + Replace Simple and Fast workflow - Lighting Lora and Multi Lora Loader

ComfyUI Workflow: Wan Animate for Flexible Pose Transfer and Video Character Replacement Wan Animate is a versatile ComfyUI workflow enabling **pose transfer from videos to static images** and **direct character replacement within videos**. It offers a straightforward approach to animating images or re-imagining video content, integrating advanced models and robust control over video processing and mask generation. **What makes Wan Animate special** - **Dual Functionality**: Seamlessly animate a static image with a video's motion or replace a character in a video. - **Optimized Performance**: Utilizes the FP8 version of Animate for efficient processing, along with specific models like umt5_xxl_fp8, Clip ViT H, and Wan 2.1 Vae. - **Multi-Lora Support**: Integrate multiple LORAs using the Power Lora Loader, pre-configured with Lightx2v and Wan Relight for enhanced output. - **Precise Pose Extraction**: Employs two DWPose Estimators for detailed face, body, and hand pose estimation from video sources. - **Intuitive Masking (Replace Workflow)**: Define character masks on the first video frame using a points editor for targeted replacement, supported by Sam2 with configurable grow and blockify options. **How it works** - **Pose Transfer Mode**: Load an image and a reference video. The video's pose is extracted via DWPose, then applied to your image through the WanAnimateToVideo sampler and KSampler to generate an animated output. - **Character Replacement Mode**: Input a video and an image of the desired character. After initial pose processing, you define a mask on the first video frame to isolate the character to be replaced. Sam2 generates the mask, which guides the sampler to integrate your new character into the video. - **Video Pre-processing**: Videos are loaded, resized, and can have their FPS adjusted and end frames capped for precise control. - **Sampling and Output**: The WanAnimateToVideo sampler manages motion coherence, especially for longer videos, while KSampler handles frame generation. Output videos are combined at the source FPS and then interpolated to 32fps by default. **Quick start in ComfyUI** - **Load Image**: Connect your input image to the designated node. - **Load Video**: Upload your reference video, adjusting FPS or capping frames as needed. - **Prompts**: Enter your positive prompt. A negative prompt is generally not required if using Lightx2v. - **Pose Transfer**: Run the workflow to animate your image with the video's motion. - **Character Replace**: For replacement, initiate the workflow to process the video, then pause at the points editor. Define your mask (green for positive, red for negative areas) and resume to generate the masked video. **Recommended settings** - **Video Dimensions**: Resize inputs to a multiple of 16 for optimal Animate model performance. - **FPS**: Start with 16 fps for general videos. Experiment with 21 fps if finer motion is desired. - **Sampling**: "Continue motion" in WanAnimateToVideo can improve long video consistency. - **Replace Masking**: Use shift + left click for positive mask points (what to replace) and shift + right click for negative points (what to exclude). - **Sam2 Masking**: Adjust 'Growmask' to expand the mask coverage and 'Blockify Mask' size (default 16) for different mask styles. Blocking can be toggled off with ctrl + b. **Why use this workflow** - **Efficient Motion Creation**: Quickly animate still images with dynamic video poses. - **Targeted Character Swaps**: Replace specific characters in videos with controlled masking. - **Advanced Model Integration**: Leverages state-of-the-art Wan Animate models for high-quality results. - **Flexible Control**: Offers fine-tuning options for video processing, LORA application, and masking parameters. - **Streamlined Process**: Once pose estimates are cached, subsequent runs for the same video are significantly faster. **Use cases** - **Animation Prototyping**: Rapidly animate character concepts from still images. - **Video Content Customization**: Replace actors or characters in existing video footage. - **Creative Video Editing**: Experiment with different characters in a scene without reshooting. - **Digital Character Development**: Visualize how new characters move within various video contexts. **Pro tips** - **First Frame Masking**: For the replace workflow, precisely mask the first frame of your video. The pose estimators do not re-estimate if dimensions remain consistent, saving time on subsequent runs. - **Pose Cache**: The initial pose estimation takes the longest. Keeping video dimensions consistent allows the workflow to reuse prior estimations, speeding up iterations. - **Lora Management**: Utilize the Power Lora Loader to experiment with different lighting or style LORAs. - **FPS Experimentation**: While 16 fps is a good starting point, test 21 fps for videos with subtle or fast movements if needed. **FAQ** - **Do I need specific hardware**: The FP8 models and general ComfyUI requirements apply. Performance will depend on your GPU. - **Can I use custom LORAs**: Yes, the Power Lora Loader allows you to load multiple custom LORAs. - **How long does a video take to process**: Initial pose estimation can take 2-3 minutes for a new video. Subsequent masking and sampling runs for the same video are much faster. - **Is the replacement workflow always perfect**: The replacement workflow is noted as somewhat experimental and might require fine-tuning for optimal quality compared to the basic pose transfer.

Screenshot of the ComfyUI promptless style transfer workflow. This template uses the Florence-2 model to apply any artistic style to an image without writing a single prompt.

This ComfyUI workflow uses Wan Animate to either transfer a pose from a video to an image or replace a video character with an image character. The pose transfer involves loading an image and video, processing the video with DWPose estimators, applying Lora models, and sampling to generate an interpolated output video. The character replacement workflow includes an additional masking step where users manually define an object to replace on the first frame, which Sam2 then uses to generate a mask for the sampler. The replacement feature is described as experimental.

Similar listings in category

A ComfyUI workflow for the Infinite Talking model, generating infinite talking videos from an audio input and a video or image input, with improved speed and quality over prior models like Multitalk.

Infinite Talking Model on Multitalk

ComfyUI Workflow: Infinite Talking Model on Multitalk The Infinite Talking model is a ComfyUI workflow designed for creating talking head videos. It allows you to generate videos where a character speaks continuously, driven by an audio input. What this workflow offers - **Infinite Talking Videos**: Produce videos where a subject appears to talk for an extended duration. - **Flexible Input**: Use either a video or an image as the visual source, combined with an audio input. - **Enhanced Performance**: Offers significant improvements in generation speed and output quality compared to previous models like Multitalk. How it works - **Audio-driven Synthesis**: An audio track guides the speech and facial movements of the character. - **Visual Source Integration**: A selected video or image provides the visual identity for the talking character. Why use this workflow - **Rapid Content Creation**: Quickly generate engaging talking head videos for various applications. - **High Fidelity Output**: Benefit from advanced quality settings for more realistic and stable results. - **Efficiency**: Experience faster processing times, making iteration and production more streamlined. Use cases - **Narrative Content**: Create narrations or presentations with a virtual presenter. - **Educational Videos**: Generate talking avatars for tutorials or lectures. - **Animated Storytelling**: Bring static images or short video clips to life with spoken dialogue.

Screenshot of the ComfyUI workflow for turning photos into Ghibli-style anime. This template uses a LoRA to capture the enchanting, hand-painted aesthetic of Hayao Miyazaki.

Wan 2.2 Light2X Image-to-Video

ComfyUI Workflow: Wan 2.2 Light2X Image-to-Video This ComfyUI workflow utilizes Wan 2.2 and the Light2X Lora to efficiently generate video from images. It focuses on rapid production and flexible output, enabling users to create dynamic clips with optimized performance. What makes this workflow efficient * **Fast Generation**: Produce videos in as few as 4 steps, significantly reducing processing time. * **Multi-Dimensional Output**: Supports various video resolutions including 480p, 720p, and custom sizes. * **High Frame Rate**: Generates 32 frames per second videos directly through integrated interpolation using FILM VFI. * **Optimized Model Loading**: Employs FP8 models to accelerate the generation process, leading directly to sampling. * **Streamlined Lora Management**: The RGTHREE Power Lora Loader allows for loading multiple loras without requiring numerous nodes. Generation speed highlights * A 5-second 480x832 video can be generated in approximately 1 minute 10 seconds after an initial run. * A 5-second 720x1280 video can be generated in about 3 minutes 30 seconds after an initial run. * This efficiency supports generating multiple short videos quickly. How to use this workflow * **Model Loading**: The workflow uses both High noise and Low noise Wan 2.2 models. FP8 models are advised for faster processing. * **Lora Integration**: Use the RGTHREE Power Lora Loader. Set the high_noise_model loader at strength 2 and the low_noise_model loader at strength 1. For other loras, load the same lora file into both loaders, with the first loader at double strength. * **VAE and Clip**: Always use the Wan 2.1 VAE. FP8 is recommended for Clip for balance between speed and quality. * **Image and Prompt**: Load your source image and select desired output dimensions. Input your text prompt and specify video length (e.g., 81 frames for 5 seconds, 160 frames for 10 seconds). * **Sampling**: Two KSamplers are utilized, one for each noise model. Light2X enables generation in just 4 steps. Euler / Simple or Euler / beta are recommended samplers. Keep CFG at 1 to maintain quality and avoid artifacts. * **Video Assembly**: The initial video output is at 16fps, which is then interpolated by FILM VFI to achieve a final 32fps video. Lora usage guidance * Wan 2.2 Loras are highly recommended, requiring both High and Low noise models from their respective Civitai pages. * Wan 2.1 Loras are compatible. Load the same lora file in both loaders, with the first (High noise) loader set at double the strength of the second. This workflow offers a streamlined and efficient method for creating high-quality, high-framerate videos from images within ComfyUI.

Screenshot of the ComfyUI workflow for animating any image into a Ghibli-style video. This template transforms a static photo into a living scene with the iconic Miyazaki aesthetic.

WAN 2.2 Text-to-Video

ComfyUI Workflow: WAN 2.2 Text-to-Video V2 Fast Workflow The **WAN 2.2 Text-to-Video V2 Fast Workflow** is an upgraded ComfyUI solution designed to generate **cinematic HD-quality videos** from a simple text prompt. This V2 edition introduces **prompt extension capabilities** and **enhanced color grading** for more vivid, high-contrast visuals. Built for **speed and quality**, it leverages WAN 2.2’s advanced motion rendering and smarter prompt handling to interpret creative ideas with greater detail, depth, and visual richness. What makes WAN 2.2 V2 special - **Prompt Extension System**: Expands your prompt intelligently for richer, more dynamic scene generation. - **Better Colors & Brightness**: Delivers improved color grading, contrast, and visual vibrancy for cinematic impact. - **Fast Rendering**: Features an optimized pipeline for quick turnaround without sacrificing detail. - **Text-to-Video Simplicity**: Input a prompt to get a smooth, cinematic video in HD. - **Enhanced Visual Depth**: Offers balanced tones, richer colors, and improved brightness. - **High-Speed Generation**: Creates professional-quality outputs in minimal time. How to get started in ComfyUI - Provide a text prompt describing your desired video. - The workflow processes the prompt to generate a cinematic HD video. Recommended settings - For best results, ULTRA PRO is recommended. - Adjust shift values if necessary for fine-tuning. - Test with other style and character loras to explore diverse visual styles. Why use this workflow - **Faster creative output**: Quickly generate AI-powered videos from text. - **Visually rich content**: Achieve more vibrant, high-contrast, and detailed cinematic visuals. - **Simplified workflow**: Create professional-quality video content with ease. Use cases - Rapid prototyping for video concepts. - Generating stylized or cinematic short clips. - Creating engaging visual content for various platforms. Conclusion The **WAN 2.2 Text-to-Video V2 Fast Workflow** offers a streamlined, high-quality method within ComfyUI for generating faster, richer, and more vibrant AI-generated videos, featuring enhanced visual depth and intelligent prompt handling.