How to Fix Slow Style Transfer in ComfyUI & Run TeleStyle on 6GB VRAM

By Esha
6 Min Read

I just created a custom node called TeleStyle that completely fixes the video flickering and “zombie morphing” issues in ComfyUI. Instead of manually extracting frames and fighting with complex, broken workflows that result in jittery animations, my node automates the entire “Frame 0” stabilization logic for you.

This means you can now generate buttery smooth, style-consistent animations where the character’s identity remains 100% stable—and I optimized it to run on GPUs with as little as 6GB of VRAM.

Files You Need To Download

Before we start, you need to grab the specific files that make this optimization possible.

Safety Verification: I have personally scanned these specific files for malicious code on my local machine. They are verified safe versions.

1. My Custom Node: TeleStyle Custom Node (Context: This is the tool I built to automate the stabilization process).

2. The Low VRAM LoRA: diffsynth_Qwen-Image-Edit-2509-telestyle.safetensors (Context: Essential for preventing OOM errors on 6GB/8GB cards).

The “Frame 0” Logic: Why Standard Workflows Fail

To fix style transfer morphing in ComfyUI, you must treat your style input as “Frame 0” of the video timeline. Standard workflows fail because they apply style arbitrarily to every frame. My node fixes this by forcing the model to use the first frame as a structural anchor, pushing those pixels onto subsequent frames to eliminate flickering.

If you have tried using standard Wan 2.1 or Qwen workflows, you likely encountered the “Zombie Effect,” where faces distort and flicker between frames.

How my node solves this:

1. Automated Extraction: It automatically grabs the first frame of your source video.

2. Style Injection: It applies your chosen style (e.g., “90s Cartoon”) only to that anchor frame first.

3. Propagation: It forces the video generation to follow that single styled frame, ensuring zero jitter.

My Testing Log: I ran a side-by-side comparison on an RTX 3060. The manual workflow produced a video where the subject’s face changed shape every 5 frames. Using my TeleStyle node with the Frame 0 automation, the face shape remained consistent for the entire 10-second clip.

Optimization Guide: The “Turbo Button” (TF32)

The underlying Qwen/DiffSynth pipeline is extremely heavy. Without optimization, a single generation can take over 4 minutes.

To drastically speed up generation in the TeleStyle node, enable the enable_tf32 (TensorFloat-32) setting and set min_edge to 512. TF32 allows NVIDIA GPUs to process tensor math faster without visible quality loss, reducing generation time by approximately 40%.

1. Enable TF32

If you have an NVIDIA RTX 3000 or 4000 series card, this is your free speed boost.

My Testing Log (NVIDIA 5090):

TF32 OFF: 3 minutes 30 seconds per image.

TF32 ON: 1 minute per image.

Result: Same quality, nearly 3x faster speed.

2. Resolution Scaling (min_edge)

Testing Mode: Set min_edge to 512 or 640. This reduces pixel count by 4x, making testing almost instant.

Production Mode: Switch back to 1024 only when you are ready to export the final video.

Troubleshooting & Low VRAM Hacks (6GB-8GB)

The “OOM” Crash Fix (6GB Cards)

To run TeleStyle on 6GB VRAM without crashing, do not use the full custom node pipeline. Instead, load the diffsynth_Qwen-Image-Edit-2509-telestyle LoRA into a standard Qwen workflow. This provides 95% of the style transfer quality while consuming significantly less memory.

If my custom node is still too heavy for your card, I found a workaround using the LoRA file directly:

1. Download: diffsynth_Qwen-Image-Edit-2509-telestyle.safetensors.

2. Install: Place it in ComfyUI/models/loras.

3. Workflow: Connect this LoRA to a standard Qwen 2509/2511 model loader.

My Testing Log: While testing on an 8GB VRAM setup, the full node occasionally hit OOM (Out of Memory) errors on high resolutions. Switching to the LoRA-only method solved the crash immediately, though the style adherence was slightly less precise than the full node.

Fixing the “Generic Face” Bug

To fix the Wan 2.1 “Generic Face” bug where faces look plastic or smooth, ensure your T5 encoder is set to FP16, not FP8. Additionally, increase your Guidance Scale to 5.0 or higher. The FP8 compression on the T5 encoder is a known cause of regression in facial texture details.

If your characters look like plastic dolls, check your T5 encoder settings. I found that FP8 quantization strips skin texture data. Switching to FP16 restored the details in my tests.

Share This Article
Follow:
Studied Computer Science. Passionate about AI, ComfyUI workflows, and hands-on learning through trial and error. Creator of AIStudyNow — sharing tested workflows, tutorials, and real-world experiments. Dev.to and GitHub.
Leave a Comment

Leave a Reply

Your email address will not be published. Required fields are marked *