"model generalization" Papers
17 papers found
Conference
Do ImageNet-trained Models Learn Shortcuts? The Impact of Frequency Shortcuts on Generalization
Shunxin Wang, Raymond Veldhuis, Nicola Strisciuglio
EvoLM: In Search of Lost Language Model Training Dynamics
Zhenting Qi, Fan Nie, Alexandre Alahi et al.
FedLWS: Federated Learning with Adaptive Layer-wise Weight Shrinking
Changlong Shi, Jinmeng Li, He Zhao et al.
Generalizing Deepfake Video Detection with Plug-and-Play: Video-Level Blending and Spatiotemporal Adapter Tuning
Zhiyuan Yan, Yandan Zhao, Shen Chen et al.
Probing Neural Combinatorial Optimization Models
Zhiqin Zhang, Yining Ma, Zhiguang Cao et al.
Rethinking the generalization of drug target affinity prediction algorithms via similarity aware evaluation
Chenbin Zhang, Zhiqiang Hu, Jiang Chuchu et al.
Sparse Training from Random Initialization: Aligning Lottery Ticket Masks using Weight Symmetry
Mohammed Adnan, Rohan Jain, Ekansh Sharma et al.
An Empirical Study of CLIP for Text-Based Person Search
Cao Min, Yang Bai, ziyin Zeng et al.
CurBench: Curriculum Learning Benchmark
Yuwei Zhou, Zirui Pan, Xin Wang et al.
Improving Neural Logic Machines via Failure Reflection
Zhiming Li, Yushi Cao, Yan Zheng et al.
Locally Estimated Global Perturbations are Better than Local Perturbations for Federated Sharpness-aware Minimization
Ziqing Fan, Shengchao Hu, Jiangchao Yao et al.
Mind the Boundary: Coreset Selection via Reconstructing the Decision Boundary
Shuo Yang, Zhe Cao, Sheng Guo et al.
On the Duality Between Sharpness-Aware Minimization and Adversarial Training
Yihao Zhang, Hangzhou He, Jingyu Zhu et al.
Quantized Prompt for Efficient Generalization of Vision-Language Models
Tianxiang Hao, Xiaohan Ding, Juexiao Feng et al.
SAH-SCI: Self-Supervised Adapter for Efficient Hyperspectral Snapshot Compressive Imaging
Haijin Zeng, Yuxi Liu, Yongyong Chen et al.
Teddy: Efficient Large-Scale Dataset Distillation via Taylor-Approximated Matching
Ruonan Yu, Songhua Liu, Jingwen Ye et al.
UNIC: Universal Classification Models via Multi-teacher Distillation
Yannis Kalantidis, Larlus Diane, Mert Bulent SARIYILDIZ et al.