Skip to content

wangqiang9/Awesome-Controllable-Video-Diffusion

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

73 Commits
Β 
Β 
Β 
Β 

Repository files navigation

Awesome-Controllable-Video-Diffusion

Awesome License: MIT

Awesome Controllable Video Generation with Diffusion Models.

Table of Contents

Pose Control

UniAnimate-DiT: Human Image Animation with Large-Scale Video Diffusion Transformer

πŸ“„ Paper | πŸ’» Code

OmniHuman-1: Rethinking the Scaling-Up of One-Stage Conditioned Human Animation Models

πŸ“„ Paper | 🌐 Project Page

EchoMimicV2: Towards Striking, Simplified, and Semi-Body Human Animation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

MikuDance: Animating Character Art with Mixed Motion Dynamics

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Diffusion as Shader: 3D-aware Video Diffusion for Versatile Video Generation Control

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

TANGO: Co-Speech Gesture Video Reenactment with Hierarchical Audio-Motion Embedding and Diffusion Interpolation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

DynamicPose: A robust image-to-video framework for portrait animation driven by pose sequences

πŸ’» Code

Alignment is All You Need: A Training-free Augmentation Strategy for Pose-guided Video Generation

πŸ“„ Paper

Follow Your Pose: Pose-Guided Text-to-Video Generation using Pose-Free Videos

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Animate Anyone: Consistent and Controllable Image-to-Video Synthesis for Character Animation

πŸ“„ Paper | 🌐 Project Page

DreaMoving: A Human Video Generation Framework based on Diffusion Models

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

MagicPose: Realistic Human Poses and Facial Expressions Retargeting with Identity-aware Diffusion

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

MagicAnimate: Temporally Consistent Human Image Animation using Diffusion Model

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Champ: Controllable and Consistent Human Image Animation with 3D Parametric Guidance

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Magic-Me: Identity-Specific Video Customized Diffusion

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

DisCo: Disentangled Control for Referring Human Dance Generation in Real World

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Human4DiT: Free-view Human Video Generation with 4D Diffusion Transformer

πŸ“„ Paper | 🌐 Project Page

MimicMotion : High-Quality Human Motion Video Generation with Confidence-aware Pose Guidance

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Follow-Your-Pose v2: Multiple-Condition Guided Character Image Animation for Stable Pose Control

πŸ“„ Paper | 🌐 Project Page

HumanVid: Demystifying Training Data for Camera-controllable Human Image Animation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

MusePose: a Pose-Driven Image-to-Video Framework for Virtual Human Generation.

πŸ’» Code

MDM: Human Motion Diffusion Model

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Audio Control

FantasyTalking: Realistic Talking Portrait Generation via Coherent Motion Synthesis

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Every Image Listens, Every Image Dances: Music-Driven Image Animation

πŸ“„ Paper

MEMO: Memory-Guided Diffusion for Expressive Talking Video Generation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Hallo2: Long-Duration and High-Resolution Audio-driven Portrait Image Animation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Co-Speech Gesture Video Generation via Motion-Decoupled Diffusion Model

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Diverse and Aligned Audio-to-Video Generation via Text-to-Video Model Adaptation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

MM-Diffusion: Learning Multi-Modal Diffusion Models for Joint Audio and Video Generation

πŸ“„ Paper | πŸ’» Code

Speech Driven Video Editing via an Audio-Conditioned Diffusion Model

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Hallo: Hierarchical Audio-Driven Visual Synthesis for Portrait Image Animation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Listen, denoise, action! Audio-driven motion synthesis with diffusion models

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

CoDi: Any-to-Any Generation via Composable Diffusion

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Generative Disco: Text-to-Video Generation for Music Visualization

πŸ“„ Paper

AADiff: Audio-Aligned Video Synthesis with Text-to-Image Diffusion

πŸ“„ Paper

EMO: Emote Portrait Alive Generating Expressive Portrait Videos with Audio2Video Diffusion Model under Weak Conditions

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Context-aware Talking Face Video Generation

πŸ“„ Paper

Expression Control

FantasyPortrait: Enhancing Multi-Character Portrait Animation with Expression-Augmented Diffusion Transformers

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

X-Portrait: Expressive Portrait Animation with Hierarchical Motion Attention

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

HelloMeme: Integrating Spatial Knitting Attentions to Embed High-Level and Fidelity-Rich Conditions in Diffusion Models

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

SkyReels-A1: Expressive Portrait Animation in Video Diffusion Transformers

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

DreamActor-M1: Holistic, Expressive and Robust Human Image Animation with Hybrid Guidance

πŸ“„ Paper | 🌐 Project Page

Follow-Your-Emoji: Fine-Controllable and Expressive Freestyle Portrait Animation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Echomimic: Lifelike audio-driven portrait animations through editable landmark conditions

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Universal Control

VACE: All-in-One Video Creation and Editing

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

ControlNeXt: Powerful and Efficient Control for Image and Video Generation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Control-A-Video: Controllable Text-to-Video Generation with Diffusion Models

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

ControlVideo: Training-free Controllable Text-to-Video Generation

πŸ“„ Paper | πŸ’» Code

TrackGo: A Flexible and Efficient Method for Controllable Video Generation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

VideoComposer: Compositional Video Synthesis with Motion Controllability

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Make-Your-Video: Customized Video Generation Using Textual and Structural Guidance

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

UniCtrl: Improving the Spatiotemporal Consistency of Text-to-Video Diffusion Models via Training-Free Unified Attention Control

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

SparseCtrl: Adding Sparse Controls to Text-to-Video Diffusion Models

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

VideoControlNet: A Motion-Guided Video-to-Video Translation Framework by Using Diffusion Model with ControlNet

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Cinemo: Consistent and Controllable Image Animation with Motion Diffusion Models

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Camera Control

MotionMaster: Training-free Camera Motion Transfer For Video Generation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

CinePreGen: Camera Controllable Video Previsualization via Engine-powered Diffusion

πŸ“„ Paper

CamViG: Camera Aware Image-to-Video Generation with Multimodal Transformers

πŸ“„ Paper

Direct-a-Video: Customized Video Generation with User-Directed Camera Movement and Object Motion

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

MotionCtrl: A Unified and Flexible Motion Controller for Video Generation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

CameraCtrl: Enabling Camera Control for Text-to-Video Generation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

VD3D: Taming Large Video Diffusion Transformers for 3D Camera Control

πŸ“„ Paper | 🌐 Project Page

Controlling Space and Time with Diffusion Models

πŸ“„ Paper | 🌐 Project Page

CamCo: Camera-Controllable 3D-Consistent Image-to-Video Generation

πŸ“„ Paper | 🌐 Project Page

Collaborative Video Diffusion: Consistent Multi-video Generation with Camera Control

πŸ“„ Paper | 🌐 Project Page

HumanVid: Demystifying Training Data for Camera-controllable Human Image Animation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Training-free Camera Control for Video Generation

πŸ“„ Paper | 🌐 Project Page

Director3D: Real-world Camera Trajectory and 3D Scene Generation from Text

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

MotionBooth: Motion-Aware Customized Text-to-Video Generation

πŸ“„ Paper | πŸ’» Code

DiffDreamer: Towards Consistent Unsupervised Single-view Scene Extrapolation with Conditional Diffusion Models

πŸ“„ Paper | 🌐 Project Page

Trajectory Control

MotionCanvas: Cinematic Shot Design with Controllable Image-to-Video Generation

πŸ“„ Paper | 🌐 Project Page

FreeTraj: Tuning-Free Trajectory Control in Video Diffusion Models

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

TrailBlazer: Trajectory Control for Diffusion-Based Video Generation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

DragNUWA: Fine-grained Control in Video Generation by Integrating Text, Image, and Trajectory

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Tora: Trajectory-oriented Diffusion Transformer for Video Generation

πŸ“„ Paper | 🌐 Project Page

Controllable Longer Image Animation with Diffusion Models

πŸ“„ Paper | 🌐 Project Page

MotionCtrl: A Unified and Flexible Motion Controller for Video Generation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

MotionBooth: Motion-Aware Customized Text-to-Video Generation

πŸ“„ Paper | πŸ’» Code

Puppet-Master: Scaling Interactive Video Generation as a Motion Prior for Part-Level Dynamics

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Direct-a-Video: Customized Video Generation with User-Directed Camera Movement and Object Motion

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Generative Image Dynamics

πŸ“„ Paper | 🌐 Project Page

Motion-Zero: Zero-Shot Moving Object Control Framework for Diffusion-Based Video Generation

πŸ“„ Paper

Video Diffusion Models are Training-free Motion Interpreter and Controlle

πŸ“„ Paper | 🌐 Project Page

Subject Control

Phantom: Subject-consistent video generation via cross-modal alignment

πŸ“„ Paper | 🌐 Project Page

Tunnel Try-on: Excavating Spatial-temporal Tunnels for High-quality Virtual Try-on in Videos

πŸ“„ Paper

Direct-a-Video: Customized Video Generation with User-Directed Camera Movement and Object Motion

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

ActAnywhere: Subject-Aware Video Background Generation

πŸ“„ Paper | 🌐 Project Page

MotionBooth: Motion-Aware Customized Text-to-Video Generation

πŸ“„ Paper | πŸ’» Code

Animate-A-Story: Storytelling with Retrieval-Augmented Video Generation

πŸ“„ Paper | πŸ’» Code

One-Shot Learning Meets Depth Diffusion in Multi-Object Videos

πŸ“„ Paper

Area Control

Boximator: Generating Rich and Controllable Motions for Video Synthesis

πŸ“„ Paper | 🌐 Project Page

Follow-Your-Click: Open-domain Regional Image Animation via Short Prompts

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

AnimateAnything: Fine-Grained Open Domain Image Animation with Motion Guidance

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Motion-I2V: Consistent and Controllable Image-to-Video Generation with Explicit Motion Modeling

πŸ“„ Paper | 🌐 Project Page

Streetscapes: Large-scale Consistent Street View Generation Using Autoregressive Video Diffusion

πŸ“„ Paper | 🌐 Project Page

Video Control

Customizing Motion in Text-to-Video Diffusion Models

πŸ“„ Paper | 🌐 Project Page

MotionClone: Training-Free Motion Cloning for Controllable Video Generation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

VMC: Video Motion Customization using Temporal Attention Adaption for Text-to-Video Diffusion Models

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Motion Inversion for Video Customization

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Brain Control

NeuroCine: Decoding Vivid Video Sequences from Human Brain Activties

πŸ“„ Paper

ID Control

FantasyID: Face Knowledge Enhanced ID-Preserving Video Generation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Concat-ID: Towards Universal Identity-Preserving Video Synthesis

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Ingredients: Blending Custom Photos with Video Diffusion Transformers

πŸ“„ Paper | πŸ’» Code

Identity-Preserving Text-to-Video Generation by Frequency Decomposition

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

VideoMaker: Zero-shot Customized Video Generation with the Inherent Force of Video Diffusion Models

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Movie Gen: A Cast of Media Foundation Models

πŸ“„ Paper

CustomCrafter: Customized Video Generation with Preserving Motion and Concept Composition Abilities

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

ID-Animator: Zero-Shot Identity-Preserving Human Video Generation

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

VideoBooth: Diffusion-based Video Generation with Image Prompts

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

Magic-Me: Identity-Specific Video Customized Diffusion

πŸ“„ Paper | 🌐 Project Page | πŸ’» Code

About

Awesome Controllable Video Generation with Diffusion Models

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published