Key Features
Real-Time V2V
Transform live video streams in real-time with minimal latency
Dynamic Prompting
Change prompts on the fly and see instant transformations in your video feed
Flexible Input Sources
Connect webcams, streaming sources, or any video input for transformation
Quick Start
Get started with StreamDiffusionV2 in seconds:Video Input
StreamDiffusionV2 is a video-to-video (V2V) model that requires a video input stream to transform. You must provide a video source (typically from a webcam or screen capture) for the model to process.Using WebcamStream Component
The easiest way to provide video input is using the built-inWebcamStream
component:
WebcamStream
component automatically:
- Requests camera permissions
- Publishes the video stream when Reactor connects
- Unpublishes when disconnected
- Handles errors and cleanup
Manual Video Publishing
For more control, you can manually publish video streams using the imperative API:The model requires video input to function. Make sure to publish a video stream after connecting and before starting generation.
Getting Started
When you first connect to the StreamDiffusionV2 model, it will be ready to receive commands but won’t start processing video until you follow the proper initialization sequence:- Set Initial Prompt: Before starting, you must set at least one prompt using
set_prompt
- Start Generation: Once you have a prompt set, call
start
to begin the video transformation - Dynamic Control: While running, you can change prompts in real-time or reset the system as needed
If you call
start
before setting an initial prompt, the command will be
ignored and the model won’t begin processing.Model Name
stream-diffusion-v2
Commands
Once connected, send commands usingreactor.sendMessage()
to control the video transformation process. Below are all available commands:
- set_prompt
- start
- reset
- set_denoising_step_list
set_prompt
Description: Set the prompt for video generation and transformation.Parameters:prompt
(string, required): The text prompt describing the desired video transformation
- Sets the active prompt that will be used to transform the incoming video stream
- Can be called at any time to change the transformation style
- Longer, more detailed prompts typically produce better results
- Changes take effect immediately if generation is already running
- Describe the desired scene: Focus on what should be present in the final video, not the transformation process
- Provide context and setting: Include details about the environment, lighting, atmosphere, and overall composition
- Specify style and mood: Describe the artistic style, color palette, lighting conditions, and emotional tone
- Be descriptive about elements: Instead of “a dog turns into a cat,” write “a cat is sitting in the scene”
- Include scene details: Mention backgrounds, objects, textures, and visual elements that should be present
- Use comprehensive descriptions: Longer, more detailed prompts typically produce better and more consistent results