Video Editing Diffusion Diffusion Models
Advanced video editing and manipulation using diffusion models, including motion control, composition, object editing, and frame interpolation. Does NOT include general video generation from text, basic inpainting tools, or video segmentation without editing capabilities.
There are 142 video editing diffusion models tracked. 1 score above 70 (verified tier). The highest-rated is hao-ai-lab/FastVideo at 78/100 with 3,232 stars. 5 of the top 10 are actively maintained.
Get all 142 projects as JSON
curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=diffusion&subcategory=video-editing-diffusion&limit=20"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
| # | Model | Score | Tier |
|---|---|---|---|
| 1 |
hao-ai-lab/FastVideo
A unified inference and post-training framework for accelerated video generation. |
|
Verified |
| 2 |
ModelTC/LightX2V
Light Image Video Generation Inference Framework |
|
Established |
| 3 |
thu-ml/TurboDiffusion
TurboDiffusion: 100–200× Acceleration for Video Diffusion Models |
|
Established |
| 4 |
PKU-YuanGroup/Helios
Helios: Real Real-Time Long Video Generation Model |
|
Established |
| 5 |
PKU-YuanGroup/MagicTime
[TPAMI 2025🔥] MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators |
|
Established |
| 6 |
Fantasy-AMAP/fantasy-talking
[ACM MM 2025] FantasyTalking: Realistic Talking Portrait Generation via... |
|
Established |
| 7 |
thu-ml/DiT-Extrapolation
Official implementation for "RIFLEx: A Free Lunch for Length Extrapolation... |
|
Established |
| 8 |
Tencent-Hunyuan/HunyuanImage-3.0
HunyuanImage-3.0: A Powerful Native Multimodal Model for Image Generation |
|
Established |
| 9 |
OpenMOSS/MOVA
MOVA: Towards Scalable and Synchronized Video–Audio Generation |
|
Established |
| 10 |
Tencent-Hunyuan/HunyuanVideo
HunyuanVideo: A Systematic Framework For Large Video Generation Model |
|
Established |
| 11 |
Lightricks/LTX-Video
Official repository for LTX-Video |
|
Established |
| 12 |
Advocate99/DiffGesture
[CVPR'2023] Taming Diffusion Models for Audio-Driven Co-Speech Gesture Generation |
|
Established |
| 13 |
Tencent/MimicMotion
High-Quality Human Motion Video Generation with Confidence-aware Pose Guidance |
|
Established |
| 14 |
PKU-YuanGroup/ConsisID
[CVPR 2025 Highlight🔥] Identity-Preserving Text-to-Video Generation by... |
|
Established |
| 15 |
YanWenKun/Hunyuan3D-2-WinPortable
📦Portable package for running Hunyuan3D 2.0/2.1 on Windows. | 混元 3D 2.0/2.1 整合包 |
|
Established |
| 16 |
YoungSeng/DiffuseStyleGesture
DiffuseStyleGesture: Stylized Audio-Driven Co-Speech Gesture Generation with... |
|
Established |
| 17 |
Tencent-Hunyuan/HunyuanCustom
HunyuanCustom: A Multimodal-Driven Architecture for Customized Video Generation |
|
Emerging |
| 18 |
Stanford-TML/EDGE
Official PyTorch Implementation of EDGE (CVPR 2023) |
|
Emerging |
| 19 |
G-U-N/Gen-L-Video
The official implementation for "Gen-L-Video: Multi-Text to Long Video... |
|
Emerging |
| 20 |
Tencent-Hunyuan/HunyuanVideo-I2V
HunyuanVideo-I2V: A Customizable Image-to-Video Model based on HunyuanVideo |
|
Emerging |
| 21 |
zli12321/FFGO-Video-Customization
Video Content Customization Using First Frame |
|
Emerging |
| 22 |
SenseTime-FVG/OpenDWM
An open source code repository of driving world models, with training,... |
|
Emerging |
| 23 |
TencentARC/GenCompositor
[ICLR 2026] GenCompositor: Generative Video Compositing with Diffusion Transformer |
|
Emerging |
| 24 |
nv-tlabs/ChronoEdit
[ICLR 2026] ChronoEdit: Towards Temporal Reasoning for Image Editing and... |
|
Emerging |
| 25 |
OpenDCAI/OpenWorldLib
Unified Codebase for Advanced World Models. |
|
Emerging |
| 26 |
QuanjianSong/UniVST
[TPAMI 2025] Official Pytorch Code of the Paper "UniVST: A Unified Framework... |
|
Emerging |
| 27 |
SandAI-org/MAGI-1
MAGI-1: Autoregressive Video Generation at Scale |
|
Emerging |
| 28 |
ChenyangQiQi/FateZero
[ICCV 2023 Oral] "FateZero: Fusing Attentions for Zero-shot Text-based Video Editing" |
|
Emerging |
| 29 |
mit-han-lab/radial-attention
[NeurIPS 2025] Radial Attention: O(nlogn) Sparse Attention with Energy Decay... |
|
Emerging |
| 30 |
omerbt/TokenFlow
Official Pytorch Implementation for "TokenFlow: Consistent Diffusion... |
|
Emerging |
| 31 |
EzioBy/Ditto
[CVPR 2026] Ditto: Scaling Instruction-Based Video Editing with a... |
|
Emerging |
| 32 |
FareedKhan-dev/text2video-from-scratch
A Straightforward, Step-by-Step Implementation of a Video Diffusion Model |
|
Emerging |
| 33 |
foivospar/NED
PyTorch implementation for NED (CVPR 2022). It can be used to manipulate the... |
|
Emerging |
| 34 |
text2cinemagraph/text2cinemagraph
Text2Cinemagraph: Text-Guided Synthesis of Eulerian Cinemagraphs [SIGGRAPH ASIA 2023] |
|
Emerging |
| 35 |
knightyxp/VideoCoF
[CVPR 2026] VideoCoF: Unified Video Editing with Temporal Reasoner |
|
Emerging |
| 36 |
ali-vilab/videocomposer
Official repo for VideoComposer: Compositional Video Synthesis with Motion... |
|
Emerging |
| 37 |
Tencent-Hunyuan/HunyuanImage-2.1
HunyuanImage-2.1: An Efficient Diffusion Model for High-Resolution (2K)... |
|
Emerging |
| 38 |
baaivision/NOVA
[ICLR 2025] Autoregressive Video Generation without Vector Quantization |
|
Emerging |
| 39 |
menyifang/MIMO
Official implementation of "MIMO: Controllable Character Video Synthesis... |
|
Emerging |
| 40 |
PangzeCheung/OmniTransfer
[CVPR 2026] OmniTransfer: All-in-one Framework for Spatio-temporal Video Transfer |
|
Emerging |
| 41 |
Vchitect/SEINE
[ICLR 2024] SEINE: Short-to-Long Video Diffusion Model for Generative... |
|
Emerging |
| 42 |
alimohammadiamirhossein/smite
Pytorch Implementation of "SMITE: Segment Me In TimE" (ICLR 2025) |
|
Emerging |
| 43 |
YBYBZhang/ControlVideo
[ICLR 2024] Official pytorch implementation of "ControlVideo: Training-free... |
|
Emerging |
| 44 |
Fantasy-AMAP/fantasy-portrait
FantasyPortrait: Enhancing Multi-Character Portrait Animation with... |
|
Emerging |
| 45 |
Zhen-Dong/Magic-Me
Codes for ID-Specific Video Customized Diffusion |
|
Emerging |
| 46 |
nihaomiao/CVPR23_LFDM
The pytorch implementation of our CVPR 2023 paper "Conditional... |
|
Emerging |
| 47 |
bytedance/X-Dyna
[CVPR 2025 Highlight] X-Dyna: Expressive Dynamic Human Image Animation |
|
Emerging |
| 48 |
CIntellifusion/GeometryForcing
[ICLR26] Official implementation of Geometry Forcing: Marrying Video... |
|
Emerging |
| 49 |
showlab/MotionDirector
[ECCV 2024 Oral] MotionDirector: Motion Customization of Text-to-Video... |
|
Emerging |
| 50 |
flymin/MagicDrive-V2
[ICCV 2025] Official implementation of the paper “MagicDrive-V2:... |
|
Emerging |
| 51 |
Kevin-thu/Epona
Official Code for Epona: Autoregressive Diffusion World Model for Autonomous... |
|
Emerging |
| 52 |
researchmm/MM-Diffusion
[CVPR'23] MM-Diffusion: Learning Multi-Modal Diffusion Models for Joint... |
|
Emerging |
| 53 |
FoundationVision/FlashVideo
[AAAI-2026]FlashVideo: Flowing Fidelity to Detail for Efficient... |
|
Emerging |
| 54 |
hustvl/MobileI2V
[ArXiv 2025] MobileI2V: Fast and High-Resolution Image-to-Video on Mobile Devices |
|
Emerging |
| 55 |
harlanhong/ACTalker
ICCV 2025 ACTalker: an end-to-end video diffusion framework for talking head... |
|
Emerging |
| 56 |
caiyuanhao1998/Open-OmniVCus
OmniVCus: Feedforward Subject-driven Video Customization with Multimodal... |
|
Emerging |
| 57 |
Kaihua-Chen/diffusion-vas
[CVPR 2025] Official code for Using Diffusion Priors for Video Amodal Segmentation |
|
Emerging |
| 58 |
RehgLab/RAVE
RAVE: Randomized Noise Shuffling for Fast and Consistent Video Editing with... |
|
Emerging |
| 59 |
LinghaoChan/HumanMAC
[ICCV-2023] Official code for work "HumanMAC: Masked Motion Completion for... |
|
Emerging |
| 60 |
JeremyCJM/DiffSHEG
[CVPR'24] DiffSHEG: A Diffusion-Based Approach for Real-Time Speech-driven... |
|
Emerging |
| 61 |
haoningwu3639/StoryGen
[CVPR 2024] Intelligent Grimm - Open-ended Visual Storytelling via Latent... |
|
Emerging |
| 62 |
Reagan1311/Mask2IV
Mask2IV: Interaction-Centric Video Generation via Mask Trajectories (AAAI 2026) |
|
Emerging |
| 63 |
lixirui142/VidToMe
Official Pytorch Implementation for "VidToMe: Video Token Merging for... |
|
Emerging |
| 64 |
TIGER-AI-Lab/ConsistI2V
ConsistI2V: Enhancing Visual Consistency for Image-to-Video Generation [TMLR 2024] |
|
Emerging |
| 65 |
UuuNyaa/blender_motion_generate_tools
motion_generate_tools is a Blender addon for generate motion using MDM:... |
|
Emerging |
| 66 |
SooLab/Free-Bloom
[NeurIPS 2023] Free-Bloom: Zero-Shot Text-to-Video Generator with LLM... |
|
Emerging |
| 67 |
Yi-Shi94/AMDM
Interactive Character Control with Auto-Regressive Motion Diffusion Models |
|
Emerging |
| 68 |
ali-vilab/VGen
Official repo for VGen: a holistic video generation ecosystem for video... |
|
Emerging |
| 69 |
songweige/content-debiased-fvd
[CVPR 2024] On the Content Bias in Fréchet Video Distance |
|
Emerging |
| 70 |
baaivision/URSA
[ICLR 2026] 🐻 Uniform Discrete Diffusion with Metric Path for Video Generation |
|
Emerging |
| 71 |
sihyun-yu/PVDM
[CVPR'23] Video Probabilistic Diffusion Models in Projected Latent Space |
|
Emerging |
| 72 |
invictus717/InteractiveVideo
InteractiveVideo: User-Centric Controllable Video Generation with... |
|
Emerging |
| 73 |
alimama-creative/M3DDM-Video-Outpainting
[ACM MM 2023] Official implementation of "Hierarchical Masked 3D Diffusion... |
|
Emerging |
| 74 |
knightyxp/VideoGrain
[ICLR 2025] VideoGrain: This repo is the official implementation of... |
|
Emerging |
| 75 |
RQ-Wu/LAMP
[CVPR 2024] | LAMP: Learn a Motion Pattern for Few-Shot Based Video Generation |
|
Emerging |
| 76 |
Da1yuqin/TCDiff
Official code for our AAAI25 oral👑 paper Harmonious Group Choreography with... |
|
Emerging |
| 77 |
vivoCameraResearch/Magic-World
official code for "magicworld: towards long-horizon stability for... |
|
Emerging |
| 78 |
vpulab/ovam
Code for the paper Open-Vocabulary Attention Maps with Token Optimization... |
|
Emerging |
| 79 |
CVL-UESTC/MVAR
ICLR 2026-MVAR: Visual Autoregressive Modeling with Scale and Spatial... |
|
Emerging |
| 80 |
DiffPoseTalk/DiffPoseTalk
DiffPoseTalk: Speech-Driven Stylistic 3D Facial Animation and Head Pose... |
|
Emerging |
| 81 |
jpthu17/GraphMotion
[NeurIPS 2023] Act As You Wish: Fine-Grained Control of Motion Diffusion... |
|
Emerging |
| 82 |
leob03/MultimodalDifMotionPred
[CVPR 2025 - HuMoGen] "MDMP: Multi-modal Diffusion for supervised Motion... |
|
Emerging |
| 83 |
SobeyMIL/TVG
code for "TVG: A Training-free Transition Video Generation Method with... |
|
Emerging |
| 84 |
Vicky0522/I2VEdit
[SIGGRAPH Asia 2024] I2VEdit: First-Frame-Guided Video Editing via... |
|
Emerging |
| 85 |
HyeonHo99/Video-Motion-Customization
VMC: Video Motion Customization using Temporal Attention Adaption for... |
|
Emerging |
| 86 |
jpthu17/DiffusionRet
[ICCV 2023] DiffusionRet: Generative Text-Video Retrieval with Diffusion Model |
|
Emerging |
| 87 |
lzz19980125/Hunyuan3D-2.1-Windows
A Windows-compatible version of Hunyuan3D-2.1 |
|
Emerging |
| 88 |
yrcong/flatten
Pytorch Implementation of FLATTEN: optical FLow-guided ATTENtion for... |
|
Emerging |
| 89 |
DuNGEOnmassster/VideoGen-of-Thought
[Neurips 2025 NextVid Workshop Oral✨] Official Implementation of... |
|
Emerging |
| 90 |
liangxuy/ReGenNet
[CVPR 2024] Official implementation of the paper "ReGenNet: Towards Human... |
|
Emerging |
| 91 |
JIA-Lab-research/Video-P2P
Video-P2P: Video Editing with Cross-attention Control |
|
Emerging |
| 92 |
diffusion-motion-transfer/diffusion-motion-transfer
Official Pytorch Implementation for "Space-Time Diffusion Features for... |
|
Emerging |
| 93 |
PhotonAISG/hunyuan-image3-finetune
Finetune HunyuanImage 3.0, a 80B unified understanding and generation model |
|
Emerging |
| 94 |
harlanhong/ICCV2023-MCNET
The official code of our ICCV2023 work: Implicit Identity Representation... |
|
Emerging |
| 95 |
MKFMIKU/vidm
[AAAI23 Oral] Official implementations of Video Implicit Diffusion Models |
|
Emerging |
| 96 |
Vchitect/VEnhancer
Official codes of VEnhancer: Generative Space-Time Enhancement for Video Generation |
|
Emerging |
| 97 |
arthur-qiu/FreeTraj
Code for FreeTraj, a tuning-free method for trajectory-controllable video generation |
|
Emerging |
| 98 |
xiefan-guo/i4vgen
[arXiv 2024] I4VGen: Image as Free Stepping Stone for Text-to-Video Generation |
|
Emerging |
| 99 |
pabloruizponce/MixerMDM
[CVPR 2025] Official Implementation of "MixerMDM: Learnable Composition of... |
|
Emerging |
| 100 |
alibaba/SRDiffusion
Accelerate Video Diffusion Inference via Sketching-Rendering Cooperation |
|
Emerging |
| 101 |
shim0114/T2V-Diffusion-Search
[NeurIPS 2025] Inference-Time Text-to-Video Alignment with Diffusion Latent... |
|
Emerging |
| 102 |
masashi-hatano/EgoH4
Official code releasse for "The Invisible EgoHand: 3D Hand Forecasting... |
|
Emerging |
| 103 |
shivangi-aneja/FaceTalk
[CVPR 2024] FaceTalk: Audio-Driven Motion Diffusion for Neural Parametric Head Models |
|
Emerging |
| 104 |
desaixie/pa_vdm
CVPRW 2025 paper Progressive Autoregressive Video Diffusion Models:... |
|
Emerging |
| 105 |
QuanjianSong/LightMotion
Official Pytorch Code of the Paper "LightMotion: A Light and Tuning-free... |
|
Experimental |
| 106 |
steve-zeyu-zhang/MotionMamba
🔥 [ECCV 2024] Motion Mamba: Efficient and Long Sequence Motion Generation |
|
Experimental |
| 107 |
SobeyMIL/MVOC
code for "MVOC:atraining-free multiple video object composition method with... |
|
Experimental |
| 108 |
aiiu-lab/MeDM
Official Pytorch Implementation of "MeDM: Mediating Image Diffusion Models... |
|
Experimental |
| 109 |
stevenlsw/physgen
PhysGen: Rigid-Body Physics-Grounded Image-to-Video Generation (ECCV 2024) |
|
Experimental |
| 110 |
FareedKhan-dev/train-text2video-scratch
This repository provides a PyTorch implementation of a video diffusion... |
|
Experimental |
| 111 |
RafailFridman/SceneScape
Official Pytorch Implementation for "SceneScape: Text-Driven Consistent... |
|
Experimental |
| 112 |
Ground-A-Video/Ground-A-Video
Ground-A-Video: Zero-shot Grounded Video Editing using Text-to-image... |
|
Experimental |
| 113 |
Gen-Verse/HermesFlow
[NeurIPS 2025] HermesFlow: Seamlessly Closing the Gap in Multimodal... |
|
Experimental |
| 114 |
jeffreychou777/GenComm
[NeurIPS 2025] Official repo for paper "Pragmatic Heterogeneous... |
|
Experimental |
| 115 |
finlay-hudson/TABE
Track Anything Behind Everything: Zero-Shot Amodal Video Object Segmentation |
|
Experimental |
| 116 |
aimagelab/VHS
[CVPR2026 Findings] VHS: Verifier on Hidden States, an efficient... |
|
Experimental |
| 117 |
ziplab/BLADE
This is the official PyTorch implementation of "BLADE: Block-Sparse... |
|
Experimental |
| 118 |
KevinDayve/VTok
Unofficial implementation of VTok (https://arxiv.org/pdf/2602.04202) |
|
Experimental |
| 119 |
wenhao728/VORTA
The code implementation of paper "VORTA: Efficient Video Diffusion via... |
|
Experimental |
| 120 |
zhang-zx/AVID
This respository contains the code for the CVPR 2024 paper AVID: Any-Length... |
|
Experimental |
| 121 |
pittisl/PhyT2V
official code repo of CVPR 2025 paper PhyT2V: LLM-Guided Iterative... |
|
Experimental |
| 122 |
k8xu/amodal
Official code for "Amodal Completion via Progressive Mixed Context... |
|
Experimental |
| 123 |
taco-group/Pulse-of-Motion
The Pulse of Motion: Measuring Physical Frame Rate from Visual Dynamics |
|
Experimental |
| 124 |
EngineeringAI-LAB/3DXTalker
Official repository for 3DXTalker: An Integrated Framework for Expressive 3D... |
|
Experimental |
| 125 |
snap-research/SF-V
This respository contains the code for the NeurIPS 2024 paper SF-V: Single... |
|
Experimental |
| 126 |
Fantasy-AMAP/fantasy-talking2
[AAAI 2026] FantasyTalking2: Timestep-Layer Adaptive Preference Optimization... |
|
Experimental |
| 127 |
Adamdad/vico
Vico: Compositional Video Generation as Flow Equalization |
|
Experimental |
| 128 |
MOSTAFA1172m/Image-text-video-I2VGENXL
A PyTorch implementation of a text-image to video diffussion model with a... |
|
Experimental |
| 129 |
Shaadalam9/traffic-pipeline
This repository contains the code and analysis for the research paper "Deep... |
|
Experimental |
| 130 |
makepixelsdance/makepixelsdance.github.io
Homepage for PixelDance. Paper -> https://arxiv.org/abs/2311.10982 |
|
Experimental |
| 131 |
DualParal-Project/DualParal
[AAAI 2026] Minute-Long Videos with Dual Parallelisms |
|
Experimental |
| 132 |
eric-ai-lab/Mojito
Official repo for the paper "Mojito: Motion Trajectory and Intensity Control... |
|
Experimental |
| 133 |
kyon317/Learned-Motion-Matching
Learned Motion Matching Implementation |
|
Experimental |
| 134 |
xie-lab-ml/IV-mixed-Sampler
[ICLR2025] IV-Mixed Sampler: Leveraging Image Diffusion Models for Enhanced... |
|
Experimental |
| 135 |
nysp78/counterfactual-video-generation
A causally faithful framework for counterfactual video generation, guided... |
|
Experimental |
| 136 |
knightyxp/EVA_Video_Edit
EVA: Zero-shot Accurate Attributes and Multi-Object Video Editing |
|
Experimental |
| 137 |
aidayang/Hunyuan3D-2.1-windows-OneClick
腾讯图片转3D模型软件混元Hunyuan3D-2.1版免安装一键启动整合包 |
|
Experimental |
| 138 |
Wang-Xiaodong1899/Long-DWM
🌟[AAAI 2026] The official repo for "LongDWM: Cross-Granularity Distillation... |
|
Experimental |
| 139 |
xiaojieli0903/GenViewPlusPlus
Official repository of "GenView++: Unifying Adaptive View Generation and... |
|
Experimental |
| 140 |
oakink/OakInk2-TaMF
[CVPR 2024] OakInk2 baseline model: Task-aware Motion Fulfillment (TaMF)... |
|
Experimental |
| 141 |
tasinislam21/FashionFlow
This model synthesises high-fidelity fashion videos from single images... |
|
Experimental |
| 142 |
synlp/T2V-Review
A collection of text-to-video generation studies. |
|
Experimental |