Exploring Best Arm with Top Reward-Cost Ratio in Stochastic Bandits
暂无分享,去创建一个
Xiaoying Gan | Jia Liu | Luoyi Fu | Zhida Qin | Haiming Jin | Hongqiu Wu | Jia Liu | Haiming Jin | Luoyi Fu | Xiaoying Gan | Zhida Qin | Hongqiu Wu
[1] Shivaram Kalyanakrishnan,et al. PAC Identification of a Bandit Arm Relative to a Reward Quantile , 2017, AAAI.
[2] Yifan Wu,et al. On Identifying Good Options under Combinatorially Structured Feedback in Finite Noisy Environments , 2015, ICML.
[3] R. Munos,et al. Best Arm Identification in Multi-Armed Bandits , 2010, COLT.
[4] Waldemar Susek,et al. Theory and Measurement of Signal-to-Noise Ratio in Continuous-Wave Noise Radar , 2018, Sensors.
[5] Jia Liu,et al. Combinatorial Sleeping Bandits with Fairness Constraints , 2019, IEEE INFOCOM 2019 - IEEE Conference on Computer Communications.
[6] Ness B. Shroff,et al. Exploring k out of Top ρ Fraction of Arms in Stochastic Bandits , 2018, ArXiv.
[7] Nenghai Yu,et al. Best Action Selection in a Stochastic Environment , 2016, AAMAS.
[8] Jonathan Scarlett,et al. Overlapping Multi-Bandit Best Arm Identification , 2019, 2019 IEEE International Symposium on Information Theory (ISIT).
[9] Jie Xu,et al. Task Replication for Vehicular Cloud: Contextual Combinatorial Bandit with Delayed Feedback , 2018, IEEE INFOCOM 2019 - IEEE Conference on Computer Communications.
[10] Alessandro Lazaric,et al. Best-Arm Identification in Linear Bandits , 2014, NIPS.
[11] T. L. Lai Andherbertrobbins. Asymptotically Efficient Adaptive Allocation Rules , 1985 .
[12] Aurélien Garivier,et al. On the Complexity of Best-Arm Identification in Multi-Armed Bandit Models , 2014, J. Mach. Learn. Res..
[13] Yingce Xia,et al. Infinitely Many-Armed Bandits with Budget Constraints , 2016, AAAI.
[14] Claire J. Tomlin,et al. Budget-Constrained Multi-Armed Bandits with Multiple Plays , 2017, AAAI.
[15] Rémi Munos,et al. Pure Exploration in Multi-armed Bandits Problems , 2009, ALT.
[16] Ambuj Tewari,et al. PAC Subset Selection in Stochastic Multi-armed Bandits , 2012, ICML.
[17] Shie Mannor,et al. PAC Bounds for Multi-armed Bandit and Markov Decision Processes , 2002, COLT.
[18] Wei Cao,et al. On Top-k Selection in Multi-Armed Bandits and Hidden Bipartite Graphs , 2015, NIPS.
[19] Rémi Munos,et al. Pure exploration in finitely-armed and continuous-armed bandits , 2011, Theor. Comput. Sci..
[20] Andrew W. Moore,et al. Hoeffding Races: Accelerating Model Selection Search for Classification and Function Approximation , 1993, NIPS.
[21] Xi Chen,et al. Optimal PAC Multiple Arm Identification with Applications to Crowdsourcing , 2014, ICML.
[22] Shie Mannor,et al. Action Elimination and Stopping Conditions for the Multi-Armed Bandit and Reinforcement Learning Problems , 2006, J. Mach. Learn. Res..
[23] Massimiliano Pontil,et al. Empirical Bernstein Bounds and Sample-Variance Penalization , 2009, COLT.
[24] Russell Greiner,et al. The Budgeted Multi-armed Bandit Problem , 2004, COLT.
[25] Ness B. Shroff,et al. Efficient Beam Alignment in Millimeter Wave Systems Using Contextual Bandits , 2017, IEEE INFOCOM 2018 - IEEE Conference on Computer Communications.
[26] Yuan Xue,et al. Pure-Exploration Bandits for Channel Selection in Mission-Critical Wireless Communications , 2018, IEEE Transactions on Vehicular Technology.
[27] Alessandro Lazaric,et al. Best Arm Identification: A Unified Approach to Fixed Budget and Fixed Confidence , 2012, NIPS.
[28] Michal Valko,et al. Simple regret for infinitely many armed bandits , 2015, ICML.
[29] Alessandro Lazaric,et al. Multi-Bandit Best Arm Identification , 2011, NIPS.
[30] Peter Stone,et al. Efficient Selection of Multiple Bandit Arms: Theory and Practice , 2010, ICML.