Poster "contrastive language-image pre-training" Papers
17 papers found
Conference
Advancing Interpretability of CLIP Representations with Concept Surrogate Model
Nhat Hoang-Xuan, Xiyuan Wei, Wanli Xing et al.
NEURIPS 2025
Advancing Myopia To Holism: Fully Contrastive Language-Image Pre-training
Haicheng Wang, Chen Ju, Weixiong Lin et al.
CVPR 2025arXiv:2412.00440
10
citations
CLIP Under the Microscope: A Fine-Grained Analysis of Multi-Object Representation
Reza Abbasi, Ali Nazari, Aminreza Sefid et al.
CVPR 2025arXiv:2502.19842
6
citations
FineLIP: Extending CLIP’s Reach via Fine-Grained Alignment with Longer Text Inputs
Mothilal Asokan, Kebin wu, Fatima Albreiki
CVPR 2025arXiv:2504.01916
15
citations
GenHancer: Imperfect Generative Models are Secretly Strong Vision-Centric Enhancers
Shijie Ma, Yuying Ge, Teng Wang et al.
ICCV 2025arXiv:2503.19480
10
citations
Harnessing Vision Foundation Models for High-Performance, Training-Free Open Vocabulary Segmentation
Yuheng Shi, Minjing Dong, Chang Xu
ICCV 2025arXiv:2411.09219
15
citations
Mitigate the Gap: Improving Cross-Modal Alignment in CLIP
Sedigheh Eslami, Gerard de Melo
ICLR 2025
15
citations
Refining CLIP's Spatial Awareness: A Visual-Centric Perspective
Congpei Qiu, Yanhao Wu, Wei Ke et al.
ICLR 2025arXiv:2504.02328
7
citations
un$^2$CLIP: Improving CLIP's Visual Detail Capturing Ability via Inverting unCLIP
Yinqi Li, Jiahe Zhao, Hong Chang et al.
NEURIPS 2025arXiv:2505.24517
1
citations
Vision-Language Model IP Protection via Prompt-based Learning
Lianyu Wang, Meng Wang, Huazhu Fu et al.
CVPR 2025arXiv:2503.02393
Alpha-CLIP: A CLIP Model Focusing on Wherever You Want
Zeyi Sun, Ye Fang, Tong Wu et al.
CVPR 2024arXiv:2312.03818
170
citations
Better Safe than Sorry: Pre-training CLIP against Targeted Data Poisoning and Backdoor Attacks
Wenhan Yang, Jingdong Gao, Baharan Mirzasoleiman
ICML 2024arXiv:2310.05862
18
citations
Bridging the Pathology Domain Gap: Efficiently Adapting CLIP for Pathology Image Analysis with Limited Labeled Data
Zhengfeng Lai, Joohi Chauhan, Brittany N. Dugger et al.
ECCV 2024
3
citations
CLIP-KD: An Empirical Study of CLIP Model Distillation
Chuanguang Yang, Zhulin An, Libo Huang et al.
CVPR 2024arXiv:2307.12732
86
citations
Gradient-based Visual Explanation for Transformer-based CLIP
Chenyang ZHAO, Kun Wang, Xingyu Zeng et al.
ICML 2024
Machine Vision Therapy: Multimodal Large Language Models Can Enhance Visual Robustness via Denoising In-Context Learning
Zhuo Huang, Chang Liu, Yinpeng Dong et al.
ICML 2024arXiv:2312.02546
23
citations
Synergy of Sight and Semantics: Visual Intention Understanding with CLIP
Qu Yang, Mang Ye, Dacheng Tao
ECCV 2024
7
citations