Skip to main content

Using VACE

VACE (Video All-in-One Creation and Editing) enables advanced video creation and editing tasks in Scope. Use reference images to define characters and styles, or control videos to guide the structure and motion of your generations.
VACE support is still experimental and the implementation is incomplete. Some features may not work as expected.

Pipeline Compatibility

VACE is supported on the following pipelines:

Wan2.1 1.3B Pipelines

  • LongLive
  • RewardForcing
  • MemFlow

Wan2.1 14B Pipeline

  • Krea Realtime
StreamDiffusion V2 also has VACE capabilities, but quality is currently limited.
Krea Realtime + VACE requires approximately 55GB of VRAM.FP8 quantization is not currently supported with VACE. Continued prompting with Krea + VACE may require resetting the cache due to cache recomputation limitations.

Supported Features

Reference-to-Video (R2V)

Use reference images to guide the character, style, and aesthetic of your generation

Video-to-Video (V2V)

Use control videos (depth, pose, scribble, optical flow) to guide the structure and motion

Animate Anything

Combine R2V + V2V: reference image defines the look, control video provides the movement

Real-time Depth

Built-in video-depth-anything preprocessor generates depth maps from source videos automatically

Built-in Preprocessors

For real-time V2V workflows, these preprocessors automatically generate control signals from your video input (webcam, screen capture, or uploaded video):
PipelineDescriptionModel Required
video-depth-anythingDepth estimation for temporally consistent depth mapsYes (~1GB VRAM)
optical-flowRAFT optical flow for motion visualizationNo (torchvision)
scribbleContour/line art extractionYes
grayGrayscale conversionNo
Select a preprocessor from the Preprocessor dropdown in the UI when using Video input mode. The preprocessor output becomes the control signal for V2V generation.
Additional preprocessors will be available via plugins in the future.

Not Yet Supported

The following features are being investigated but not currently available:
  • Multiple reference images for R2V
  • Masked video-to-video (MV2V) for inpainting, outpainting, and video extension
  • Complex tasks like Swap Anything, Reference Anything, Move Anything, Expand Anything

Enabling VACE

Before using any VACE features, make sure VACE is enabled in your pipeline settings.
1

Open Settings

Click the Settings panel in the Scope interface.
2

Enable VACE

Toggle VACE to On.
VACE toggle in Settings panel

Reference-to-Video (R2V)

Use a reference image to guide the character, style, or aesthetic of your generation. The model will try to maintain consistency with the reference throughout the video.
1

Add a reference image

In the Settings panel, find Reference Images and click Add Image.
Add Image button under Reference Images
2

Select your image

Use the media picker to either:
  • Upload a new image
  • Select from your previously uploaded assets
Media picker for selecting reference images
3

Verify the reference

You should see a preview of your selected reference image in the panel.
Reference image preview
Only a single reference image is supported at this time. Multi-reference support is planned for a future release.

Video-to-Video (V2V)

Use a control video to guide the structure and motion of your generation. Control videos can be depth maps, pose estimations, scribbles, or optical flow visualizations.
1

Set input mode to Video

Under Input & Controls, set Input Mode to Video.
2

Upload a control video

Upload your control video (e.g., a depth map or pose estimation video).
V2V input settings
3

Generate

The output will follow the structure of your control video while applying the style from your prompt.

Example Control Video

A pose estimation video can be used to transfer motion to AI-generated characters:

Animate Anything

Combine Reference-to-Video and Video-to-Video for the best of both worlds:
  • Reference image → Defines the character, style, and aesthetic
  • Control video → Provides the structure and motion
This is powerful for animating still images or transferring motion to custom characters.

Example Workflow

1

Add a reference image

Upload an image of your character or style reference.
Reference image for Animate Anything
2

Add a control video

Upload a pose or depth video that contains the motion you want.
3

Generate

The output combines the look from your reference with the motion from your control video.

Example Results

Basic Animate Anything: With LoRA enhancement: For even better character consistency, combine Animate Anything with a LoRA. This example uses the Arcane Jinx LoRA:

Tips for Best Results

Use high-quality reference images with clear subjects. The model works best when the reference has good lighting and a clean background.
Match your control video resolution to your pipeline’s output resolution for best structural accuracy.
For character consistency, pair VACE with a relevant LoRA. This helps maintain style and identity across the generation.
If quality degrades during long sessions with Krea + VACE, try resetting the cache to restore output quality.

API Usage

For programmatic control of VACE features, see the API reference:

VACE API Reference

Upload reference images, set context scale, and combine with control videos via WebRTC

See Also