1. Upload photo
2. Choose Model
3. Add Script
Trusted by teams
Overview
Zero‑shot video‑to‑video lip sync that preserves a speaker’s style while matching new audio. Built for editing dialogue or dubbing across live‑action, animation, and AI‑generated humans without retraining.
Highlights
- Zero‑shot editing with no actor training required.
- Preserves unique speaking style and cadence.
- Works with live‑action, animation, and AI‑generated characters.
- High‑resolution workflows up to 4K.
Quick Specifications
Best for
Inputs & Outputs
UGC ad re‑dub
Swap a new hook while preserving the original footage.
Capabilities
Zero‑shot editing
- No per‑speaker training required.
- Preserves original performance style.
- Works across live‑action and animation.
Dubbing workflows
- Swap dialogue quickly for new scripts.
- Maintain timing alignment to the original cut.
Use Cases
UGC variations
Rotate new scripts without reshoots.
Explainers
Keep visuals, change narration fast.
Creator content
Ship updates with the same host.
Applications
Marketing refresh
Update scripts without reshooting.
Creator content
Publish new hooks with the same footage.
Localization prep
Create a clean base for language versions.
Best Practices
- 1Use clear, studio‑quality audio for the target voice.
- 2Keep the face large and well‑lit for best mouth detail.
- 3Match the emotional tone of the original performance.
Frequently Asked Questions
Do I need to train on the speaker first?
No. Lipsync‑2 is zero‑shot, so it can edit any speaker without training.
What kinds of footage does it support?
It works on live‑action video, animation, and AI‑generated humans.
What inputs are required?
Provide a source video plus target audio (or a script + voice) via the API.
