ComfyUITemplates.com
Discover free ready-made ComfyUI templates for AI workflows.
Wan Animate Pose transfer + Replace Simple and Fast workflow - Lighting Lora and Multi Lora Loader
Wan Animate is a ComfyUI workflow for pose transfer from video to image and character replacement within video. Features multi-LoRA support, DWPose estimation, Sam2 masking, and 32fps interpolated output.
ComfyUI Workflow: Wan Animate – Pose Transfer and Video Character Replacement
Wan Animate is a versatile ComfyUI workflow enabling pose transfer from videos to static images and direct character replacement within videos. It offers a straightforward approach to animating images or re-imagining video content, integrating advanced models and robust control over video processing and mask generation.
What Makes Wan Animate Special
- Dual functionality: Seamlessly animate a static image with a video's motion, or replace a character in a video with an image character.
- Optimized performance: Utilizes the FP8 version of Wan Animate for efficient processing, along with umt5_xxl_fp8, CLIP ViT H, and Wan 2.1 VAE models.
- Multi-LoRA support: Integrate multiple LoRAs using the Power LoRA Loader, pre-configured with Lightx2v and Wan Relight for enhanced output.
- Precise pose extraction: Employs two DWPose Estimators for detailed face, body, and hand pose estimation from video sources.
- Intuitive masking (replace workflow): Define character masks on the first video frame using a points editor for targeted replacement, supported by Sam2 with configurable grow and blockify options.
How It Works
Pose Transfer Mode Load an image and a reference video. The video's pose is extracted via DWPose, then applied to your image through the WanAnimateToVideo sampler and KSampler to generate an animated output.
Character Replacement Mode Input a video and an image of the desired character. After initial pose processing, define a mask on the first video frame to isolate the character to be replaced. Sam2 generates the mask, which guides the sampler to integrate your new character into the video.
Video Pre-processing and Output Videos are loaded, resized, and can have their FPS adjusted and end frames capped for precise control. Output videos are combined at the source FPS and then interpolated to 32fps by default.
Quick Start in ComfyUI
- Load image: Connect your input image to the designated node.
- Load video: Upload your reference video, adjusting FPS or capping frames as needed.
- Prompts: Enter your positive prompt. A negative prompt is generally not required if using Lightx2v.
- Pose transfer: Run the workflow to animate your image with the video's motion.
- Character replace: Initiate the workflow to process the video, then pause at the points editor. Define your mask (green for positive, red for negative areas) and resume to generate the masked video.
Recommended Settings
- Video dimensions: Resize inputs to a multiple of 16 for optimal Wan Animate model performance.
- FPS: Start with 16fps for general videos. Try 21fps if finer motion is desired.
- Sampling: "Continue motion" in WanAnimateToVideo can improve consistency for longer videos.
- Replace masking: Use Shift + Left Click for positive mask points (what to replace) and Shift + Right Click for negative points (what to exclude).
- Sam2 masking: Adjust Growmask to expand coverage and Blockify Mask size (default 16) for different mask styles. Toggle blocking off with Ctrl + B.
Why Use This Workflow
- Efficient motion creation: Quickly animate still images with dynamic video poses.
- Targeted character swaps: Replace specific characters in videos with controlled masking.
- Advanced model integration: Leverages state-of-the-art Wan Animate models for high-quality results.
- Flexible control: Fine-tune video processing, LoRA application, and masking parameters.
- Streamlined process: Once pose estimates are cached, subsequent runs for the same video are significantly faster.
Use Cases
- Animating character concepts from still images for rapid prototyping
- Replacing actors or characters in existing video footage
- Creative video editing without reshooting scenes
- Visualizing how new characters move within various video contexts
Pro Tips
- First frame masking: For the replace workflow, precisely mask the first frame. Pose estimators do not re-estimate if dimensions remain consistent, saving time on subsequent runs.
- Pose cache: Initial pose estimation takes longest. Keeping video dimensions consistent allows the workflow to reuse prior estimations, speeding up iterations.
- LoRA management: Use the Power LoRA Loader to experiment with different lighting or style LoRAs.
- FPS experimentation: While 16fps is a good starting point, test 21fps for videos with subtle or fast movements.
FAQ
- Do I need specific hardware? The FP8 models and standard ComfyUI requirements apply. Performance depends on your GPU.
- Can I use custom LoRAs? Yes, the Power LoRA Loader supports multiple custom LoRAs.
- How long does processing take? Initial pose estimation can take 2–3 minutes for a new video. Subsequent runs for the same video are much faster.
- Is the replacement workflow always perfect? The replacement workflow is experimental and may require fine-tuning for optimal quality compared to the basic pose transfer.