Krea Realtime Video Pipeline
Krea Realtime Video is a streaming pipeline and autoregressive video diffusion model from Krea. The model is trained using Self-Forcing on Wan2.1 14b.At a Glance
| Base Model | Wan2.1 14B |
| Estimated VRAM | ~32GB (40GB+ recommended) |
| Training | Self-Forcing |
| LoRA Support | 14B LoRAs |
| VACE Support | Yes |
| T2V / V2V | Yes / Limited* |
Examples
The following examples include timeline JSON files with the prompts used so you can try them as well. A GPU with >40GB VRAM (e.g. H100, RTX 6000 Pro) is recommended for these examples since they use a higher resolution.Flower Bloom
Abstract Shape
A >= 32 GB VRAM GPU (eg RTX 5090) is recommended for these examples which have lower VRAM requirements due to the lower resolution.Flower Bloom (Low Resolution)
Acceleration
The pipeline uses different attention kernels to accelerate inference depending on the hardware used:- SageAttention 2 is used on all GPUs except for Hopper GPUs (eg H100). If you run into video quality issues (which some folks have reported while using SageAttention) you can restart the server with
DISABLE_SAGEATTENTION=1(egDISABLE_SAGEATTENTION=1 uv run daydream-scope) to fallback to Flash Attention 2. - Flash Attention 2 is the fallback if SageAttention 2 is disabled.
- Flash Attention 3 is used on Hopper GPUs (eg H100).
Resolution
The generation will be faster for smaller resolutions resulting in smoother video. The visual quality will be better at higher resolutions (eg 832x480 and larger), but you may need a more powerful GPU in order to achieve a higher FPS.Seed
The seed parameter in the UI can be used to reproduce generations. If you like the generation for a certain seed value and sequence of prompts you can re-use that value later with those same prompts to reproduce the generation.Prompting
Subject and Background/Setting Anchors The model works better if you include a subject (who/what) and background/setting (where) in each prompt. If you want continuity in the next scene then you can continue referencing the same subject and/or background/setting. For example:Offline Generation
A test script can be used for offline generation. If the model weights are not downloaded yet:output.mp4 file in the krea_realtime_video directory.
See Also
Other Pipelines
StreamDiffusion V2
Real-time streaming from the original StreamDiffusion creators
LongLive
Smooth prompt transitions and extended generation from Nvidia
RewardForcing
Reward-matched training for improved output quality
MemFlow
Memory bank for long-context consistency