Paper "zero-shot learning" Papers
25 papers found
Conference
Advancing Comprehensive Aesthetic Insight with Multi-Scale Text-Guided Self-Supervised Learning
Yuti Liu, Shice Liu, Junyuan Gao et al.
Attend and Enrich: Enhanced Visual Prompt for Zero-Shot Learning
Man Liu, Huihui Bai, Feng Li et al.
EMPLACE: Self-Supervised Urban Scene Change Detection
Tim Alpherts, Sennay Ghebreab, Nanne van Noord
Holistic Semantic Representation for Navigational Trajectory Generation
Ji Cao, Tongya Zheng, Qinghong Guo et al.
Limitations in Employing Natural Language Supervision for Sensor-Based Human Activity Recognition - And Ways to Overcome Them
Harish Haresamudram, Apoorva Beedu, Mashfiqui Rabbi et al.
MeRino: Entropy-Driven Design for Generative Language Models on IoT Devices
Youpeng Zhao, Ming Lin, Huadong Tang et al.
Standing on the Shoulders of Giants: Reprogramming Visual-Language Model for General Deepfake Detection
Kaiqing Lin, Yuzhen Lin, Weixiang Li et al.
Universal Features Guided Zero-Shot Category-Level Object Pose Estimation
Wentian Qu, Chenyu Meng, Heng Li et al.
ZeroMamba: Exploring Visual State Space Model for Zero-Shot Learning
Wenjin Hou, Dingjie Fu, Kun Li et al.
BDIQA: A New Dataset for Video Question Answering to Explore Cognitive Reasoning through Theory of Mind
Yuanyuan Mao, Xin Lin, Qin Ni et al.
Chinese Spelling Correction as Rephrasing Language Model
Linfeng Liu, Hongqiu Wu, Hai Zhao
Commonsense for Zero-Shot Natural Language Video Localization
Meghana Holla, Ismini Lourentzou
Context-I2W: Mapping Images to Context-Dependent Words for Accurate Zero-Shot Composed Image Retrieval
Yuanmin Tang, Jing Yu, Keke Gai et al.
CORECODE: A Common Sense Annotated Dialogue Dataset with Benchmark Tasks for Chinese Large Language Models
Dan Shi, Chaobin You, Jian-Tao Huang et al.
Data Adaptive Traceback for Vision-Language Foundation Models in Image Classification
Long-Fei Li, Peng Zhao, Zhi-Hua Zhou
Data-Free Generalized Zero-Shot Learning
Bowen Tang, Jing Zhang, Yan Long et al.
GroundVLP: Harnessing Zero-Shot Visual Grounding from Vision-Language Pre-training and Open-Vocabulary Object Detection
Haozhan Shen, Tiancheng Zhao, Mingwei Zhu et al.
Image Captioning with Multi-Context Synthetic Data
Feipeng Ma, Y. Zhou, Fengyun Rao et al.
InstructDoc: A Dataset for Zero
Shot Generalization of Visual Document Understanding with Instructions - Ryota Tanaka, Taichi Iki, Kyosuke Nishida et al.
Interactive Visual Task Learning for Robots
Weiwei Gu, Anant Sah, N. Gopalan
Prompting Segmentation with Sound Is Generalizable Audio-Visual Source Localizer
Yaoting Wang, Liu Weisong, Guangyao Li et al.
StyleSinger: Style Transfer for Out
of-Domain Singing Voice Synthesis
Task Contamination: Language Models May Not Be Few-Shot Anymore
Changmao Li, Jeffrey Flanigan
Wikiformer: Pre-training with Structured Information of Wikipedia for Ad-Hoc Retrieval
Weihang Su, Qingyao Ai, Xiangsheng Li et al.
Zero-1-to-3: Domain-Level Zero-Shot Cognitive Diagnosis via One Batch of Early-Bird Students towards Three Diagnostic Objectives
Weibo Gao, Qi Liu, Hao Wang et al.