"temporal consistency" Papers

88 papers found • Page 1 of 2

3D StreetUnveiler with Semantic-aware 2DGS - a simple baseline

Jingwei Xu, Yikai Wang, Yiqun Zhao et al.

ICLR 2025oralarXiv:2405.18416
4
citations

ARLON: Boosting Diffusion Transformers with Autoregressive Models for Long Video Generation

Zongyi Li, Shujie HU, Shujie LIU et al.

ICLR 2025oralarXiv:2410.20502
28
citations

A Unified Solution to Video Fusion: From Multi-Frame Learning to Benchmarking

Zixiang Zhao, Haowen Bai, Bingxin Ke et al.

NEURIPS 2025oralarXiv:2505.19858
7
citations

Buffer Anytime: Zero-Shot Video Depth and Normal from Image Priors

Zhengfei Kuang, Tianyuan Zhang, Kai Zhang et al.

CVPR 2025arXiv:2411.17249
4
citations

Coherent 3D Portrait Video Reconstruction via Triplane Fusion

Shengze Wang, Xueting Li, Chao Liu et al.

CVPR 2025arXiv:2405.00794
2
citations

Consistent and Controllable Image Animation with Motion Diffusion Models

Xin Ma, Yaohui Wang, Gengyun Jia et al.

CVPR 2025arXiv:2407.15642
13
citations

Consistent Time-of-Flight Depth Denoising via Graph-Informed Geometric Attention

Weida Wang, Changyong He, Jin Zeng et al.

ICCV 2025arXiv:2506.23542

Ctrl-Adapter: An Efficient and Versatile Framework for Adapting Diverse Controls to Any Diffusion Model

Han Lin, Jaemin Cho, Abhay Zala et al.

ICLR 2025oralarXiv:2404.09967
50
citations

CTRL-D: Controllable Dynamic 3D Scene Editing with Personalized 2D Diffusion

Kai He, Chin-Hsuan Wu, Igor Gilitschenski

CVPR 2025arXiv:2412.01792
5
citations

Depth Any Video with Scalable Synthetic Data

Honghui Yang, Di Huang, Wei Yin et al.

ICLR 2025oralarXiv:2410.10815
46
citations

Diffusion$^2$: Dynamic 3D Content Generation via Score Composition of Video and Multi-view Diffusion Models

Zeyu Yang, Zijie Pan, Chun Gu et al.

ICLR 2025oralarXiv:2404.02148
20
citations

Dirichlet-Constrained Variational Codebook Learning for Temporally Coherent Video Face Restoration

Baoyou Chen, Ce Liu, Weihao Yuan et al.

ICCV 2025highlightarXiv:2506.13355

DIVE: Taming DINO for Subject-Driven Video Editing

Yi Huang, Wei Xiong, He Zhang et al.

ICCV 2025arXiv:2412.03347
9
citations

EG4D: Explicit Generation of 4D Object without Score Distillation

Qi Sun, Zhiyang Guo, Ziyu Wan et al.

ICLR 2025oralarXiv:2405.18132
40
citations

EIDT-V: Exploiting Intersections in Diffusion Trajectories for Model-Agnostic, Zero-Shot, Training-Free Text-to-Video Generation

Diljeet Jagpal, Xi Chen, Vinay P. Namboodiri

CVPR 2025arXiv:2504.06861
2
citations

Emergent Temporal Correspondences from Video Diffusion Transformers

Jisu Nam, Soowon Son, Dahyun Chung et al.

NEURIPS 2025oralarXiv:2506.17220
11
citations

Event-Enhanced Blurry Video Super-Resolution

Dachun Kai, Yueyi Zhang, Jin Wang et al.

AAAI 2025paperarXiv:2504.13042
6
citations

FlashDepth: Real-time Streaming Video Depth Estimation at 2K Resolution

Gene Chou, Wenqi Xian, Guandao Yang et al.

ICCV 2025highlightarXiv:2504.07093
6
citations

FlowMo: Variance-Based Flow Guidance for Coherent Motion in Video Generation

Ariel Shaulov, Itay Hazan, Lior Wolf et al.

NEURIPS 2025oralarXiv:2506.01144
8
citations

FramePainter: Endowing Interactive Image Editing with Video Diffusion Priors

Yabo Zhang, xinpeng zhou, Yihan Zeng et al.

ICCV 2025arXiv:2501.08225
12
citations

From Prompt to Progression: Taming Video Diffusion Models for Seamless Attribute Transition

Ling Lo, Kelvin Chan, Wen-Huang Cheng et al.

ICCV 2025arXiv:2509.19690
1
citations

Generating, Fast and Slow: Scalable Parallel Video Generation with Video Interface Networks

Bhishma Dedhia, David Bourgin, Krishna Kumar Singh et al.

ICCV 2025arXiv:2503.17539
1
citations

Geometry-guided Online 3D Video Synthesis with Multi-View Temporal Consistency

Hyunho Ha, Lei Xiao, Christian Richardt et al.

CVPR 2025arXiv:2505.18932

Glad: A Streaming Scene Generator for Autonomous Driving

Bin Xie, Yingfei Liu, Tiancai Wang et al.

ICLR 2025oralarXiv:2503.00045
11
citations

High Temporal Consistency through Semantic Similarity Propagation in Semi-Supervised Video Semantic Segmentation for Autonomous Flight

Cédric Vincent, Taehyoung Kim, Henri Meeß

CVPR 2025arXiv:2503.15676
3
citations

HumanMM: Global Human Motion Recovery from Multi-shot Videos

Yuhong Zhang, Guanlin Wu, Ling-Hao Chen et al.

CVPR 2025arXiv:2503.07597
3
citations

Image as a World: Generating Interactive World from Single Image via Panoramic Video Generation

Dongnan Gui, Xun Guo, Wengang Zhou et al.

NEURIPS 2025oral
1
citations

Incremental Sequence Classification with Temporal Consistency

Lucas Maystre, Gabriel Barello, Tudor Berariu et al.

NEURIPS 2025oralarXiv:2505.16548

Infinite-Resolution Integral Noise Warping for Diffusion Models

Yitong Deng, Winnie Lin, Lingxiao Li et al.

ICLR 2025oralarXiv:2411.01212
4
citations

LongDiff: Training-Free Long Video Generation in One Go

Zhuoling Li, Hossein Rahmani, Qiuhong Ke et al.

CVPR 2025arXiv:2503.18150
5
citations

Lux Post Facto: Learning Portrait Performance Relighting with Conditional Video Diffusion and a Hybrid Dataset

Yiqun Mei, Mingming He, Li Ma et al.

CVPR 2025arXiv:2503.14485
17
citations

MATE: Motion-Augmented Temporal Consistency for Event-based Point Tracking

Han Han, Wei Zhai, Yang Cao et al.

ICCV 2025arXiv:2412.01300

MPG-SAM 2: Adapting SAM 2 with Mask Priors and Global Context for Referring Video Object Segmentation

Fu Rong, Meng Lan, Qian Zhang et al.

ICCV 2025arXiv:2501.13667
3
citations

OCK: Unsupervised Dynamic Video Prediction with Object-Centric Kinematics

YeonJi Song, Jaein Kim, Suhyung Choi et al.

ICCV 2025arXiv:2404.18423

Ouroboros-Diffusion: Exploring Consistent Content Generation in Tuning-free Long Video Diffusion

Jingyuan Chen, Fuchen Long, Jie An et al.

AAAI 2025paperarXiv:2501.09019
10
citations

PPMStereo: Pick-and-Play Memory Construction for Consistent Dynamic Stereo Matching

WANG Yun, Qiaole Dong, Yongjian Zhang et al.

NEURIPS 2025oralarXiv:2510.20178

Rationalizing and Augmenting Dynamic Graph Neural Networks

Guibin Zhang, Yiyan Qi, Ziyang Cheng et al.

ICLR 2025oral

ReCon-GS: Continuum-Preserved Guassian Streaming for Fast and Compact Reconstruction of Dynamic Scenes

Jiaye Fu, Qiankun Gao, Chengxiang Wen et al.

NEURIPS 2025oral

SAM4D: Segment Anything in Camera and LiDAR Streams

Jianyun Xu, Song Wang, Ziqian Ni et al.

ICCV 2025arXiv:2506.21547
4
citations

ST$^2$360D: Spatial-to-Temporal Consistency for Training-free 360 Monocular Depth Estimation

Zidong Cao, Jinjing Zhu, Hao Ai et al.

NEURIPS 2025oral

Stable Virtual Camera: Generative View Synthesis with Diffusion Models

Jensen Zhou, Hang Gao, Vikram Voleti et al.

ICCV 2025arXiv:2503.14489
87
citations

STAR: Spatial-Temporal Augmentation with Text-to-Video Models for Real-World Video Super-Resolution

Rui Xie, Yinhong Liu, Penghao Zhou et al.

ICCV 2025arXiv:2501.02976
27
citations

STDD: Spatio-Temporal Dual Diffusion for Video Generation

Shuaizhen Yao, Xiaoya Zhang, Xin Liu et al.

CVPR 2025
2
citations

Temporal Rate Reduction Clustering for Human Motion Segmentation

Xianghan Meng, Zhengyu Tong, Zhiyuan Huang et al.

ICCV 2025arXiv:2506.21249

TokensGen: Harnessing Condensed Tokens for Long Video Generation

Wenqi Ouyang, Zeqi Xiao, Danni Yang et al.

ICCV 2025arXiv:2507.15728
3
citations

Towards Generalizable Scene Change Detection

Jae-Woo KIM, Ue-Hwan Kim

CVPR 2025arXiv:2409.06214
8
citations

Tracktention: Leveraging Point Tracking to Attend Videos Faster and Better

Zihang Lai, Andrea Vedaldi

CVPR 2025highlightarXiv:2503.19904
4
citations

Training-Free Generation of Temporally Consistent Rewards from VLMs

Yinuo Zhao, Jiale Yuan, Zhiyuan Xu et al.

ICCV 2025arXiv:2507.04789
2
citations

VACE: All-in-One Video Creation and Editing

Zeyinzi Jiang, Zhen Han, Chaojie Mao et al.

ICCV 2025arXiv:2503.07598
181
citations

VideoGigaGAN: Towards Detail-rich Video Super-Resolution

Yiran Xu, Taesung Park, Richard Zhang et al.

CVPR 2025arXiv:2404.12388
27
citations
PreviousNext