Poster "multi-armed bandits" Papers
17 papers found
Conference
Breaking the $\log(1/\Delta_2)$ Barrier: Better Batched Best Arm Identification with Adaptive Grids
Tianyuan Jin, Qin Zhang, Dongruo Zhou
ICLR 2025
Constrained Feedback Learning for Non-Stationary Multi-Armed Bandits
Shaoang Li, Jian Li
NEURIPS 2025arXiv:2509.15073
Efficient Top-m Data Values Identification for Data Selection
Xiaoqiang Lin, Xinyi Xu, See-Kiong Ng et al.
ICLR 2025
LASeR: Learning to Adaptively Select Reward Models with Multi-Arm Bandits
Duy Nguyen, Archiki Prasad, Elias Stengel-Eskin et al.
NEURIPS 2025arXiv:2410.01735
6
citations
Learning Across the Gap: Hybrid Multi-armed Bandits with Heterogeneous Offline and Online Data
Qijia He, Minghan Wang, Xutong Liu et al.
NEURIPS 2025
On Speeding Up Language Model Evaluation
Jin Zhou, Christian Belardi, Ruihan Wu et al.
ICLR 2025arXiv:2407.06172
6
citations
Pairwise Elimination with Instance-Dependent Guarantees for Bandits with Cost Subsidy
Ishank Juneja, Carlee Joe-Wong, Osman Yagan
ICLR 2025arXiv:2501.10290
Pareto Optimal Risk-Agnostic Distributional Bandits with Heavy-Tail Rewards
Kyungjae Lee, Dohyeong Kim, Taehyun Cho et al.
NEURIPS 2025
Revisiting Follow-the-Perturbed-Leader with Unbounded Perturbations in Bandit Problems
Jongyeong Lee, Junya Honda, Shinji Ito et al.
NEURIPS 2025arXiv:2508.18604
2
citations
Satisficing Regret Minimization in Bandits
Qing Feng, Tianyi Ma, Ruihao Zhu
ICLR 2025arXiv:2406.06802
1
citations
Tightening Regret Lower and Upper Bounds in Restless Rising Bandits
Cristiano Migali, Marco Mussi, Gianmarco Genalti et al.
NEURIPS 2025
uniINF: Best-of-Both-Worlds Algorithm for Parameter-Free Heavy-Tailed MABs
Yu Chen, Jiatai Huang, Yan Dai et al.
ICLR 2025arXiv:2410.03284
6
citations
Causal Bandits: The Pareto Optimal Frontier of Adaptivity, a Reduction to Linear Bandits, and Limitations around Unknown Marginals
Ziyi Liu, Idan Attias, Daniel Roy
ICML 2024arXiv:2407.00950
2
citations
Factored-Reward Bandits with Intermediate Observations
Marco Mussi, Simone Drago, Marcello Restelli et al.
ICML 2024
Federated Combinatorial Multi-Agent Multi-Armed Bandits
Fares Fourati, Mohamed-Slim Alouini, Vaneet Aggarwal
ICML 2024arXiv:2405.05950
8
citations
Incentivized Learning in Principal-Agent Bandit Games
Antoine Scheid, Daniil Tiapkin, Etienne Boursier et al.
ICML 2024arXiv:2403.03811
17
citations
On Interpolating Experts and Multi-Armed Bandits
Houshuang Chen, Yuchen He, Chihao Zhang
ICML 2024arXiv:2307.07264
5
citations