Using VACE
VACE (Video All-In-One Creation and Editing) enables guiding generation using reference images and control videos. This guide shows how to use VACE for style transfer, character consistency, and video transformation.Overview
VACE allows you to:- Condition generation on reference images (style, character, scene)
- Use control videos to preserve structure and motion
- Control the influence strength of visual conditioning
Prerequisites
- Pipeline loaded with VACE enabled (default):
- For reference images: Images uploaded or available in the assets directory
- For control videos: A video input source (webcam, screen capture, or file)
Uploading Reference Images
Upload images via the assets API:Setting Reference Images
Via Initial Parameters
Set reference images when starting the WebRTC connection:Via Data Channel
Update reference images during streaming:VACE Parameters
| Parameter | Type | Range | Default | Description |
|---|---|---|---|---|
vace_ref_images | array | - | [] | List of reference image paths |
vace_context_scale | float | 0.0-2.0 | 1.0 | Visual conditioning strength |
Context Scale
Thevace_context_scale controls how strongly reference images influence generation:
- 0.0: No reference influence (pure text-to-video)
- 0.5: Subtle influence, more creative freedom
- 1.0: Balanced influence (default)
- 1.5: Strong influence, closer to reference
- 2.0: Maximum influence, may reduce diversity
Using Control Videos
When VACE is enabled, you can send a control video to guide generation while preserving the structure and motion of your input. This works the same way as sending video in regular video-to-video mode.Sending Control Video
Set up a WebRTC connection with video input, just as you would for Send and Receive Video:Combining Control Video with Reference Images
You can use both control video and reference images together for maximum control:- Use the control video for motion and structure
- Use reference images for style, character appearance, or scene elements