oguzhanercan 's Collections Video Generation
updated
DynamicScaler: Seamless and Scalable Video Generation for Panoramic
Scenes
Paper
• 2412.11100
• Published
• 7
LinGen: Towards High-Resolution Minute-Length Text-to-Video Generation
with Linear Computational Complexity
Paper
• 2412.09856
• Published
• 11
DisPose: Disentangling Pose Guidance for Controllable Human Image
Animation
Paper
• 2412.09349
• Published
• 8
MEMO: Memory-Guided Diffusion for Expressive Talking Video Generation
Paper
• 2412.04448
• Published
• 10
SynCamMaster: Synchronizing Multi-Camera Video Generation from Diverse
Viewpoints
Paper
• 2412.07760
• Published
• 55
Track4Gen: Teaching Video Diffusion Models to Track Points Improves
Video Generation
Paper
• 2412.06016
• Published
• 20
Video Creation by Demonstration
Paper
• 2412.09551
• Published
• 9
You See it, You Got it: Learning 3D Creation on Pose-Free Videos at
Scale
Paper
• 2412.06699
• Published
• 12
Mind the Time: Temporally-Controlled Multi-Event Video Generation
Paper
• 2412.05263
• Published
• 10
LeviTor: 3D Trajectory Oriented Image-to-Video Synthesis
Paper
• 2412.15214
• Published
• 15
Autoregressive Video Generation without Vector Quantization
Paper
• 2412.14169
• Published
• 14
DiTCtrl: Exploring Attention Control in Multi-Modal Diffusion
Transformer for Tuning-Free Multi-Prompt Longer Video Generation
Paper
• 2412.18597
• Published
• 20
TransPixar: Advancing Text-to-Video Generation with Transparency
Paper
• 2501.03006
• Published
• 25
Diffusion Adversarial Post-Training for One-Step Video Generation
Paper
• 2501.08316
• Published
• 36
RepVideo: Rethinking Cross-Layer Representation for Video Generation
Paper
• 2501.08994
• Published
• 15
EMO2: End-Effector Guided Audio-Driven Avatar Video Generation
Paper
• 2501.10687
• Published
• 15
Go-with-the-Flow: Motion-Controllable Video Diffusion Models Using
Real-Time Warped Noise
Paper
• 2501.08331
• Published
• 20
OmniHuman-1: Rethinking the Scaling-Up of One-Stage Conditioned Human
Animation Models
Paper
• 2502.01061
• Published
• 223
VideoJAM: Joint Appearance-Motion Representations for Enhanced Motion
Generation in Video Models
Paper
• 2502.02492
• Published
• 66
Goku: Flow Based Video Generative Foundation Models
Paper
• 2502.04896
• Published
• 106
FlashVideo:Flowing Fidelity to Detail for Efficient High-Resolution
Video Generation
Paper
• 2502.05179
• Published
• 24
Magic 1-For-1: Generating One Minute Video Clips within One Minute
Paper
• 2502.07701
• Published
• 36
Step-Video-T2V Technical Report: The Practice, Challenges, and Future of
Video Foundation Model
Paper
• 2502.10248
• Published
• 57
Mobius: Text to Seamless Looping Video Generation via Latent Shift
Paper
• 2502.20307
• Published
• 18
Open-Sora 2.0: Training a Commercial-Level Video Generation Model in
$200k
Paper
• 2503.09642
• Published
• 20
DreamRelation: Relation-Centric Video Customization
Paper
• 2503.07602
• Published
• 14
Make Your Training Flexible: Towards Deployment-Efficient Video Models
Paper
• 2503.14237
• Published
• 5
Video-T1: Test-Time Scaling for Video Generation
Paper
• 2503.18942
• Published
• 90
AnimeGamer: Infinite Anime Life Simulation with Next Game State
Prediction
Paper
• 2504.01014
• Published
• 70
Towards Physically Plausible Video Generation via VLM Planning
Paper
• 2503.23368
• Published
• 40
MoCha: Towards Movie-Grade Talking Character Synthesis
Paper
• 2503.23307
• Published
• 139
DiTaiListener: Controllable High Fidelity Listener Video Generation with
Diffusion
Paper
• 2504.04010
• Published
• 9
Seaweed-7B: Cost-Effective Training of Video Generation Foundation Model
Paper
• 2504.08685
• Published
• 130
Packing Input Frame Context in Next-Frame Prediction Models for Video
Generation
Paper
• 2504.12626
• Published
• 51
RealisDance-DiT: Simple yet Strong Baseline towards Controllable
Character Animation in the Wild
Paper
• 2504.14977
• Published
• 10
ReVision: High-Quality, Low-Cost Video Generation with Explicit 3D
Physics Modeling for Complex Motion and Interaction
Paper
• 2504.21855
• Published
• 13
Seedance 1.0: Exploring the Boundaries of Video Generation Models
Paper
• 2506.09113
• Published
• 107
Self Forcing: Bridging the Train-Test Gap in Autoregressive Video
Diffusion
Paper
• 2506.08009
• Published
• 30
FilMaster: Bridging Cinematic Principles and Generative AI for Automated
Film Generation
Paper
• 2506.18899
• Published
• 6
VMoBA: Mixture-of-Block Attention for Video Diffusion Models
Paper
• 2506.23858
• Published
• 31
A Survey on Long-Video Storytelling Generation: Architectures,
Consistency, and Cinematic Quality
Paper
• 2507.07202
• Published
• 25
UniVA: Universal Video Agent towards Open-Source Next-Generation Video Generalist
Paper
• 2511.08521
• Published
• 38
LongCat-Video Technical Report
Paper
• 2510.22200
• Published
• 33
Kandinsky 5.0: A Family of Foundation Models for Image and Video Generation
Paper
• 2511.14993
• Published
• 231
STARFlow-V: End-to-End Video Generative Modeling with Normalizing Flow
Paper
• 2511.20462
• Published
• 32
HunyuanVideo 1.5 Technical Report
Paper
• 2511.18870
• Published
• 28
SemanticGen: Video Generation in Semantic Space
Paper
• 2512.20619
• Published
• 93
LTX-2: Efficient Joint Audio-Visual Foundation Model
Paper
• 2601.03233
• Published
• 154
MOVA: Towards Scalable and Synchronized Video-Audio Generation
Paper
• 2602.08794
• Published
• 154