Poster "multi-armed bandits" Papers

17 papers found

Breaking the $\log(1/\Delta_2)$ Barrier: Better Batched Best Arm Identification with Adaptive Grids

Tianyuan Jin, Qin Zhang, Dongruo Zhou

ICLR 2025

Constrained Feedback Learning for Non-Stationary Multi-Armed Bandits

Shaoang Li, Jian Li

NEURIPS 2025arXiv:2509.15073

Efficient Top-m Data Values Identification for Data Selection

Xiaoqiang Lin, Xinyi Xu, See-Kiong Ng et al.

ICLR 2025

LASeR: Learning to Adaptively Select Reward Models with Multi-Arm Bandits

Duy Nguyen, Archiki Prasad, Elias Stengel-Eskin et al.

NEURIPS 2025arXiv:2410.01735
6
citations

Learning Across the Gap: Hybrid Multi-armed Bandits with Heterogeneous Offline and Online Data

Qijia He, Minghan Wang, Xutong Liu et al.

NEURIPS 2025

On Speeding Up Language Model Evaluation

Jin Zhou, Christian Belardi, Ruihan Wu et al.

ICLR 2025arXiv:2407.06172
6
citations

Pairwise Elimination with Instance-Dependent Guarantees for Bandits with Cost Subsidy

Ishank Juneja, Carlee Joe-Wong, Osman Yagan

ICLR 2025arXiv:2501.10290

Pareto Optimal Risk-Agnostic Distributional Bandits with Heavy-Tail Rewards

Kyungjae Lee, Dohyeong Kim, Taehyun Cho et al.

NEURIPS 2025

Revisiting Follow-the-Perturbed-Leader with Unbounded Perturbations in Bandit Problems

Jongyeong Lee, Junya Honda, Shinji Ito et al.

NEURIPS 2025arXiv:2508.18604
2
citations

Satisficing Regret Minimization in Bandits

Qing Feng, Tianyi Ma, Ruihao Zhu

ICLR 2025arXiv:2406.06802
1
citations

Tightening Regret Lower and Upper Bounds in Restless Rising Bandits

Cristiano Migali, Marco Mussi, Gianmarco Genalti et al.

NEURIPS 2025

uniINF: Best-of-Both-Worlds Algorithm for Parameter-Free Heavy-Tailed MABs

Yu Chen, Jiatai Huang, Yan Dai et al.

ICLR 2025arXiv:2410.03284
6
citations

Causal Bandits: The Pareto Optimal Frontier of Adaptivity, a Reduction to Linear Bandits, and Limitations around Unknown Marginals

Ziyi Liu, Idan Attias, Daniel Roy

ICML 2024arXiv:2407.00950
2
citations

Factored-Reward Bandits with Intermediate Observations

Marco Mussi, Simone Drago, Marcello Restelli et al.

ICML 2024

Federated Combinatorial Multi-Agent Multi-Armed Bandits

Fares Fourati, Mohamed-Slim Alouini, Vaneet Aggarwal

ICML 2024arXiv:2405.05950
8
citations

Incentivized Learning in Principal-Agent Bandit Games

Antoine Scheid, Daniil Tiapkin, Etienne Boursier et al.

ICML 2024arXiv:2403.03811
17
citations

On Interpolating Experts and Multi-Armed Bandits

Houshuang Chen, Yuchen He, Chihao Zhang

ICML 2024arXiv:2307.07264
5
citations