"communication overhead" Papers
6 papers found
Conference
Accelerating Parallel Diffusion Model Serving with Residual Compression
Jiajun Luo, Yicheng Xiao, Jianru Xu et al.
NEURIPS 2025oralarXiv:2507.17511
Block-Diagonal LoRA for Eliminating Communication Overhead in Tensor Parallel LoRA Serving
Xinyu Wang, Jonas M. Kübler, Kailash Budhathoki et al.
NEURIPS 2025arXiv:2510.23346
Communication-Efficient Diffusion Denoising Parallelization via Reuse-then-Predict Mechanism
Kunyun Wang, Bohan Li, Kai Yu et al.
NEURIPS 2025arXiv:2505.14741
1
citations
DUO: No Compromise to Accuracy Degradation
Jinda Jia, Cong Xie, Hanlin Lu et al.
NEURIPS 2025
FedCross: Intertemporal Federated Learning Under Evolutionary Games
Jianfeng Lu, Ying Zhang, Riheng Jia et al.
AAAI 2025paperarXiv:2412.16968
First Attentions Last: Better Exploiting First Attentions for Efficient Parallel Training
Gyudong Kim, Hyukju Na, Jin Kim et al.
NEURIPS 2025