WAN 2.2 Animate: ComfyUI, InfiniteTalk & Uni3C (GGUF Q4, Low VRAM)

Esha
By Esha
7 Min Read

WAN 2.2 Animate is a character-focused video model. It takes motion and expression from a driver video and applies it to your subject while keeping identity steady, so the face feels stable and the head and body follow the reference. Under the hood, Wan 2.2 uses a Mixture-of-Experts setup that improves quality across denoising steps, which is why the motion transfer looks clean when you set it up right.

There are ready builds for Animate on Hugging Face, including quantized GGUF variants that load well in ComfyUI with a GGUF loader. Repack notes also list where to place the text encoder and VAE for a working stack.

I kept the Workflow simple. One graph. You load your video. You load your subject image. You add the mask with cricle green points. You pick models. You run. The swap looks natural. Faces stay same. Motion reads right.

Files you need

WAN 2.2 Animate models

InfinityTalk for lip sync

Uni3C for camera motion

LoRAs used in this flow

The workflow layout

Background masking

  1. Load your video in the Video section.
  2. In Resolution Master, pick a supported size for the model and your GPU. It helps you stay inside safe memory.
  3. Open the Point Editor. Use green points for what you want in the mask and red points for what you want to exclude. Keep only the object you plan to replace.
  4. Prefer points for speed. If you like layer masks, use SegmentAnything Ultra v3 style nodes in Segments v3. For this tutorial we stay with SEM to segment nodes.

Face motion from a driver video

  1. If you do not like the expression in your first video, load another short face video as the driver.
  2. Plug the Image hook from that driver clip into the Face Image section group.
  3. If you want to stick to the first video’s face motion, connect the first video back to that hook.

Lip sync

  1. Enable InfinityTalk with WAN 2.2 Animate when you want stronger lip sync. It is an audio driven model that drives mouth and subtle head motion.
  2. If you run Animate in GGUF, pair it with the Wan 2.1 InfiniteTalk GGUF build to match the loader.

Camera motion

  1. Turn on WanVideo Uni3C when you want camera pan or subtle angle. It transfers camera movement from a reference clip. In ComfyUI you load the Uni3C ControlNet file and connect the control video.
  2. It is still evolving. Results vary across builds.

Model section

  1. Pick WAN 2.2 Animate FP8 for quality. If your GPU is tight on VRAM, try Q4 GGUF.
  2. Load LoRAs
    • Relighting to fix color tone and lighting match
    • LightX 4-step for speed and motion stability
    • Pusa v1 if you need more motion
    • HPS if you want human-preference tilt toward realism
  3. Text encoder and VAE
    • Use the WAN text encoder and VAE that your Animate build expects. See the model page for exact pairs

Guitar test with Image swap

Reference image: Gollum cut out with Remove Background.

Video: a woman playing guitar.

Masking: green points on the woman shape, red points on the guitar and background edges.

Run.

The DW pose node reads face and body points and passes motion through, so hands keep strum beats while you swap the subject.

Q4 GGUF vs FP8

First try on Q4 took about 28 seconds to show a first result. I had the frame cap stuck at 37 by mistake. I fixed it to 153 frames and it split across two runs. Total was around one minute on my side.

FP8 Model
GGUF Q4

Quality matched the FP8 sample. Motion looked better once the frame cap was corrected.

If your card is small, Q4 is a real option. The QuantStack GGUF repo lists Q2 to Q8 variants so you can move up or down by VRAM.

change the face motion with a driver clip

I loaded a new face driver video where the head bobs up and down.

Swapped the hook to this driver.

On the first run the motion only came partway.

I ran again with a tiny prompt nudge and it followed the driver 100 percent.

This is the quick trick. When expression is wrong, load another driver and reconnect the hook.

Lip sync with InfinityTalk

Turn on the InfinityTalk group.

Video: a woman singing. Reference image: a woman portrait.

First I only check the mask. Then I enable models and pick the InfinityTalk build.

If Animate is GGUF, select the InfinityTalk GGUF too.

Run. Lip sync locks in. You can zoom close and it still looks blended. Necklace, clothes and face match the reference look.

Test Uni3C for camera motion

Enable the Uni3C group.

Load a short clip that has camera angle changes.

I used 720 by 720.

On my side the result was mostly static with a small drift. This is expected sometimes. Try different driver clips. The Uni3C paper and loaders are new and moving fast.

wITH Uni3C
Without Uni3C

FAQ

Can I use this on a very small GPU
Yes. Try the Q4 GGUF builds and smaller resolutions.

Share This Article
Follow:
Studied Computer Science. Passionate about AI, ComfyUI workflows, and hands-on learning through trial and error. Creator of AIStudyNow — sharing tested workflows, tutorials, and real-world experiments.
Leave a Comment

Leave a Reply

Your email address will not be published. Required fields are marked *