"diffusion models" Papers

1,011 papers found • Page 11 of 21

SliderSpace: Decomposing the Visual Capabilities of Diffusion Models

Rohit Gandikota, Zongze Wu, Richard Zhang et al.

ICCV 2025arXiv:2502.01639
10
citations

SlowFast-VGen: Slow-Fast Learning for Action-Driven Long Video Generation

Yining Hong, Beide Liu, Maxine Wu et al.

ICLR 2025oralarXiv:2410.23277
19
citations

SMITE: Segment Me In TimE

Amirhossein Alimohammadi, Sauradip Nag, Saeid Asgari et al.

ICLR 2025arXiv:2410.18538
7
citations

Solving Video Inverse Problems Using Image Diffusion Models

Taesung Kwon, Jong Chul YE

ICLR 2025oralarXiv:2409.02574
10
citations

Spatial Transport Optimization by Repositioning Attention Map for Training-Free Text-to-Image Synthesis

Woojung Han, Yeonkyung Lee, Chanyoung Kim et al.

CVPR 2025arXiv:2503.22168
5
citations

Spectral Motion Alignment for Video Motion Transfer Using Diffusion Models

Geon Yeong Park, Hyeonho Jeong, Sang Wan Lee et al.

AAAI 2025paperarXiv:2403.15249
19
citations

Spiral: Semantic-Aware Progressive LiDAR Scene Generation and Understanding

Dekai Zhu, Yixuan Hu, Youquan Liu et al.

NEURIPS 2025arXiv:2505.22643
5
citations

StableAnimator: High-Quality Identity-Preserving Human Image Animation

Shuyuan Tu, Zhen Xing, Xintong Han et al.

CVPR 2025arXiv:2411.17697
64
citations

StableCodec: Taming One-Step Diffusion for Extreme Image Compression

Tianyu Zhang, Xin Luo, Li Li et al.

ICCV 2025arXiv:2506.21977
7
citations

Stable Flow: Vital Layers for Training-Free Image Editing

Omri Avrahami, Or Patashnik, Ohad Fried et al.

CVPR 2025arXiv:2411.14430
60
citations

Stable Part Diffusion 4D: Multi-View RGB and Kinematic Parts Video Generation

Hao Zhang, Chun-Han Yao, Simon Donné et al.

NEURIPS 2025oralarXiv:2509.10687
3
citations

Stable Score Distillation

Haiming Zhu, Yangyang Xu, Chenshu Xu et al.

ICCV 2025arXiv:2507.09168
2
citations

Stable Virtual Camera: Generative View Synthesis with Diffusion Models

Jensen Zhou, Hang Gao, Vikram Voleti et al.

ICCV 2025arXiv:2503.14489
87
citations

STAR: Spatial-Temporal Augmentation with Text-to-Video Models for Real-World Video Super-Resolution

Rui Xie, Yinhong Liu, Penghao Zhou et al.

ICCV 2025arXiv:2501.02976
27
citations

State-Covering Trajectory Stitching for Diffusion Planners

Kyowoon Lee, Jaesik Choi

NEURIPS 2025oralarXiv:2506.00895
4
citations

STDD: Spatio-Temporal Dual Diffusion for Video Generation

Shuaizhen Yao, Xiaoya Zhang, Xin Liu et al.

CVPR 2025
2
citations

STITCH-OPE: Trajectory Stitching with Guided Diffusion for Off-Policy Evaluation

Hossein Goli, Michael Gimelfarb, Nathan de Lara et al.

NEURIPS 2025spotlightarXiv:2505.20781
2
citations

Straight-Line Diffusion Model for Efficient 3D Molecular Generation

Yuyan Ni, Shikun Feng, Haohan Chi et al.

NEURIPS 2025arXiv:2503.02918
7
citations

Straightness of Rectified Flow: A Theoretical Insight into Wasserstein Convergence

Saptarshi Roy, Vansh Bansal, Purnamrita Sarkar et al.

ICLR 2025

StreamDiffusion: A Pipeline-level Solution for Real-Time Interactive Generation

Akio Kodaira, Chenfeng Xu, Toshiki Hazama et al.

ICCV 2025arXiv:2312.12491
64
citations

StreamingT2V: Consistent, Dynamic, and Extendable Long Video Generation from Text

Roberto Henschel, Levon Khachatryan, Hayk Poghosyan et al.

CVPR 2025arXiv:2403.14773
164
citations

Stretching Each Dollar: Diffusion Training from Scratch on a Micro-Budget

Vikash Sehwag, Xianghao Kong, Jingtao Li et al.

CVPR 2025arXiv:2407.15811
26
citations

Structure-Guided Diffusion Models for High-Fidelity Portrait Shadow Removal

wanchang Yu, Qing Zhang, Rongjia Zheng et al.

ICCV 2025arXiv:2507.04692

Structure Language Models for Protein Conformation Generation

Jiarui Lu, Xiaoyin Chen, Stephen Lu et al.

ICLR 2025arXiv:2410.18403
21
citations

StyleKeeper: Prevent Content Leakage using Negative Visual Query Guidance

Jaeseok Jeong, Junho Kim, Youngjung Uh et al.

ICCV 2025arXiv:2510.06827
2
citations

StyleSSP: Sampling StartPoint Enhancement for Training-free Diffusion-based Method for Style Transfer

ruojun xu, Weijie Xi, Xiaodi Wang et al.

CVPR 2025highlightarXiv:2501.11319
6
citations

SummDiff: Generative Modeling of Video Summarization with Diffusion

Kwanseok Kim, Jaehoon Hahm, Sumin Kim et al.

ICCV 2025highlightarXiv:2510.08458

SuperPC: A Single Diffusion Model for Point Cloud Completion, Upsampling, Denoising, and Colorization

Yi Du, Zhipeng Zhao, Shaoshu Su et al.

CVPR 2025arXiv:2503.14558
9
citations

SVDQuant: Absorbing Outliers by Low-Rank Component for 4-Bit Diffusion Models

Muyang Li, Yujun Lin, Zhekai Zhang et al.

ICLR 2025arXiv:2411.05007
98
citations

SwiftTry: Fast and Consistent Video Virtual Try-On with Diffusion Models

Hung Nguyen, Quang Qui-Vinh Nguyen, Khoi Nguyen et al.

AAAI 2025paperarXiv:2412.10178
7
citations

SymmCD: Symmetry-Preserving Crystal Generation with Diffusion Models

Daniel Levy, Siba Smarak Panigrahi, Sékou-Oumar Kaba et al.

ICLR 2025arXiv:2502.03638
41
citations

SyncVP: Joint Diffusion for Synchronous Multi-Modal Video Prediction

Enrico Pallotta, Sina Mokhtarzadeh Azar, Shuai Li et al.

CVPR 2025arXiv:2503.18933

Synthetic Data is an Elegant GIFT for Continual Vision-Language Models

Bin Wu, Wuxuan Shi, Jinqiao Wang et al.

CVPR 2025arXiv:2503.04229
15
citations

Synthetic-powered predictive inference

Meshi Bashari, Roy Maor Lotan, Yonghoon Lee et al.

NEURIPS 2025arXiv:2505.13432
4
citations

T2ICount: Enhancing Cross-modal Understanding for Zero-Shot Counting

Yifei Qian, Zhongliang Guo, Bowen Deng et al.

CVPR 2025highlightarXiv:2502.20625
9
citations

T2SMark: Balancing Robustness and Diversity in Noise-as-Watermark for Diffusion Models

Jindong Yang, Han Fang, Weiming Zhang et al.

NEURIPS 2025arXiv:2510.22366
3
citations

T2V-OptJail: Discrete Prompt Optimization for Text-to-Video Jailbreak Attacks

Jiayang Liu, Siyuan Liang, Shiqian Zhao et al.

NEURIPS 2025arXiv:2505.06679
6
citations

TACO: Taming Diffusion for in-the-wild Video Amodal Completion

Ruijie Lu, Yixin Chen, Yu Liu et al.

ICCV 2025arXiv:2503.12049
10
citations

TADA: Improved Diffusion Sampling with Training-free Augmented DynAmics

Tianrong Chen, Huangjie Zheng, David Berthelot et al.

NEURIPS 2025arXiv:2506.21757
1
citations

TaxaDiffusion: Progressively Trained Diffusion Model for Fine-Grained Species Generation

Amin Karimi Monsefi, Mridul Khurana, Rajiv Ramnath et al.

ICCV 2025arXiv:2506.01923
3
citations

TCAQ-DM: Timestep-Channel Adaptive Quantization for Diffusion Models

Haocheng Huang, Jiaxin Chen, Jinyang Guo et al.

AAAI 2025paperarXiv:2412.16700
3
citations

TCFG: Tangential Damping Classifier-free Guidance

Mingi Kwon, Shin seong Kim, Jaeseok Jeong et al.

CVPR 2025arXiv:2503.18137
6
citations

TD-Paint: Faster Diffusion Inpainting Through Time-Aware Pixel Conditioning

Tsiry MAYET, Pourya Shamsolmoali, Simon Bernard et al.

ICLR 2025arXiv:2410.09306

TeethGenerator: A two-stage framework for paired pre- and post-orthodontic 3D dental data generation

Changsong Lei, Yaqian Liang, Shaofeng Wang et al.

ICCV 2025arXiv:2507.04685
1
citations

TEncDM: Understanding the Properties of the Diffusion Model in the Space of Language Model Encodings

Alexander Shabalin, Viacheslav Meshchaninov, Egor Chimbulatov et al.

AAAI 2025paperarXiv:2402.19097
9
citations

Text2Data: Low-Resource Data Generation with Textual Control

Shiyu Wang, Yihao Feng, Tian Lan et al.

AAAI 2025paperarXiv:2402.10941

Text2Relight: Creative Portrait Relighting with Text Guidance

Junuk Cha, Mengwei Ren, Krishna Kumar Singh et al.

AAAI 2025paperarXiv:2412.13734
6
citations

TexTailor: Customized Text-aligned Texturing via Effective Resampling

Suin Lee, DAE SHIK KIM

ICLR 2025arXiv:2506.10612

Text-Aware Real-World Image Super-Resolution via Diffusion Model with Joint Segmentation Decoders

Qiming Hu, Linlong Fan, Yiyan Luo et al.

NEURIPS 2025arXiv:2506.04641
2
citations

Text-to-Image Rectified Flow as Plug-and-Play Priors

Xiaofeng Yang, Cheng Chen, xulei yang et al.

ICLR 2025arXiv:2406.03293
23
citations