"model fine-tuning" Papers
23 papers found
Conference
$\mathcal{X}^2$-DFD: A framework for e$\mathcal{X}$plainable and e$\mathcal{X}$tendable Deepfake Detection
Yize Chen, Zhiyuan Yan, Guangliang Cheng et al.
A Large-scale Training Paradigm for Graph Generative Models
Yu Wang, Ryan Rossi, Namyong Park et al.
An OpenMind for 3D Medical Vision Self-supervised Learning
Tassilo Wald, Constantin Ulrich, Jonathan Suprijadi et al.
CAMEx: Curvature-aware Merging of Experts
Dung Viet Nguyen, Minh Nguyen, Luc Nguyen et al.
Can LLMs Separate Instructions From Data? And What Do We Even Mean By That?
Egor Zverev, Sahar Abdelnabi, Soroush Tabesh et al.
Can LLMs Solve Longer Math Word Problems Better?
Xin Xu, Tong Xiao, Zitong Chao et al.
Detail-Preserving Latent Diffusion for Stable Shadow Removal
Jiamin Xu, Yuxin Zheng, Zelong Li et al.
Differentially Private Federated Low Rank Adaptation Beyond Fixed-Matrix
Ming Wen, Jiaqi Zhu, Yuedong Xu et al.
Escaping the SpuriVerse: Can Large Vision-Language Models Generalize Beyond Seen Spurious Correlations?
Yiwei Yang, Chung Peng Lee, Shangbin Feng et al.
Flick: Empowering Federated Learning with Commonsense Knowledge
Ran Zhu, Mingkun Yang, Shiqiang Wang et al.
FlowerTune: A Cross-Domain Benchmark for Federated Fine-Tuning of Large Language Models
Yan Gao, Massimo R. Scamarcia, Javier Fernandez-Marques et al.
Machine Unlearning via Simulated Oracle Matching
Kristian G Georgiev, Roy Rinberg, Sam Park et al.
Persistent Pre-training Poisoning of LLMs
Yiming Zhang, Javier Rando, Ivan Evtimov et al.
Singular Value Scaling: Efficient Generative Model Compression via Pruned Weights Refinement
Hyeonjin Kim, Jaejun Yoo
Small Batch Size Training for Language Models: When Vanilla SGD Works, and Why Gradient Accumulation is Wasteful
Martin Marek, Sanae Lotfi, Aditya Somasundaram et al.
Spend Wisely: Maximizing Post-Training Gains in Iterative Synthetic Data Bootstrapping
Pu Yang, Yunzhen Feng, Ziyuan Chen et al.
Trade-offs in Image Generation: How Do Different Dimensions Interact?
Sicheng Zhang, Binzhu Xie, Zhonghao Yan et al.
$\texttt{MoE-RBench}$: Towards Building Reliable Language Models with Sparse Mixture-of-Experts
Guanjie Chen, Xinyu Zhao, Tianlong Chen et al.
Adversarial Feature Map Pruning for Backdoor
Dong HUANG, Qingwen Bu
Language Models as Science Tutors
Alexis Chevalier, Jiayi Geng, Alexander Wettig et al.
MGit: A Model Versioning and Management System
Wei Hao, Daniel Mendoza, Rafael Mendes et al.
Neural Lineage
Runpeng Yu, Xinchao Wang
PoseGen: Learning to Generate 3D Human Pose Dataset with NeRF
Mohsen Gholami, Rabab Ward, Z. Jane Wang