Video-to-Video
Video-to-video models take one or more videos as input and generate new videos as output. They can enhance quality, interpolate frames, modify styles, or create new motion dynamics, enabling creative applications, video production, and research.


About Video-to-Video
Use Cases
Video Style Transfer
Apply artistic or cinematic styles to a video while preserving motion and structure. For example, convert real footage into anime, painting, or film-like visuals.
Frame Interpolation
Generate intermediate frames to make videos smoother or convert 30 FPS videos to 60 FPS. This improves motion flow and enables realistic slow-motion playback.
Video Super-Resolution
Enhance low-resolution videos into high-definition outputs with preserved detail and sharpness. Ideal for restoring old footage or improving video quality.
Motion Transfer
Transfer the motion from a source video to another subject while maintaining identity and environment. This enables realistic animation or gesture replication.
Video Editing & Synthesis
Add, remove, or modify objects in videos while keeping lighting and motion consistent. Perfect for visual effects, object replacement, and content-aware editing.
Temporal Modification
Change a video’s overall time or environmental conditions, such as day to night or summer to winter. These models preserve motion dynamics and lighting continuity.
Virtual Try-on
Simulate clothing changes or outfit fitting in videos while keeping the person’s motion and identity intact. Useful for digital fashion and e-commerce applications.
Inference
Below is an example demonstrating how to use Lucy-Edit-Dev to perform video costume editing, changing a character’s clothing while maintaining identity and motion consistency. Lucy-Edit-Dev is trained on paired video edits, captioned videos, and extended image–text datasets.
!pip install torch diffusers
import torch
from PIL import Image
from diffusers import AutoencoderKLWan, LucyEditPipeline
from diffusers.utils import export_to_video, load_video
url = "https://d2drjpuinn46lb.cloudfront.net/painter_original_edit.mp4"
prompt = "Change the apron and blouse to a classic clown costume: satin polka-dot jumpsuit in bright primary colors, ruffled white collar, oversized pom-pom buttons, white gloves, oversized red shoes, red foam nose; soft window light from left, eye-level medium shot, natural folds and fabric highlights."
negative_prompt = ""
num_frames = 81
height = 480
width = 832
def convert_video(video: List[Image.Image]) -> List[Image.Image]:
video = load_video(url)[:num_frames]
video = [video[i].resize((width, height)) for i in range(num_frames)]
return video
video = load_video(url, convert_method=convert_video)
model_id = "decart-ai/Lucy-Edit-Dev"
vae = AutoencoderKLWan.from_pretrained(model_id, subfolder="vae", torch_dtype=torch.float32)
pipe = LucyEditPipeline.from_pretrained(model_id, vae=vae, torch_dtype=torch.bfloat16)
pipe.to("cuda")
output = pipe(
prompt=prompt,
video=video,
negative_prompt=negative_prompt,
height=480,
width=832,
num_frames=81,
guidance_scale=5.0
).frames[0]
export_to_video(output, "output.mp4", fps=24)
For more inference examples, check out the model cards on Hugging Face, where you can try the provided example code.
Useful Resources
You can read more about the datasets, model architectures, and open-source implementations in the following repositories:
- Lumen - Official implementation of Lumen for text-guided video editing.
- VIRES - Implementation for sketch- and text-guided video instance repainting.
- ECCV2022-RIFE: Video Frame Interpolation - Real-time video frame interpolation via intermediate flow estimation.
- StableVSR: Enhancing Perceptual Quality in Video - Super-resolution method to enhance perceptual video quality.
Compatible libraries
No example widget is defined for this task.
Note Contribute by proposing a widget for this task !
Note Model for editing outfits, character, and scenery in videos.
Note Framework that uses 3D mesh proxies for precise, consistent video editing.
Note Model for generating physics-aware videos from input videos and control conditions.
Note A model to upscale videos at input, designed for seamless use with ComfyUI.
Note Dataset with detailed annotations for training and benchmarking video instance editing.
Note Dataset to evaluate models on long video generation and understanding.
Note Collection of 104 demo videos from the SeedVR/SeedVR2 series showcasing model outputs.
Note Interactive demo space for Lucy-Edit-Dev video editing.
Note Demo space for SeedVR2-3B showcasing video upscaling and restoration.
No example metric is defined for this task.
Note Contribute by proposing a metric for this task !