暂无分享,去创建一个
[1] Lihong Li,et al. An Empirical Evaluation of Thompson Sampling , 2011, NIPS.
[2] Zhiyuan Liu,et al. Incentivized Exploration for Multi-Armed Bandits under Reward Drift , 2020, AAAI.
[3] Wei Chu,et al. A contextual-bandit approach to personalized news article recommendation , 2010, WWW '10.
[4] T. L. Lai Andherbertrobbins. Asymptotically Efficient Adaptive Allocation Rules , 2022 .
[5] Alessandro Lazaric,et al. Linear Thompson Sampling Revisited , 2016, AISTATS.
[6] Theja Tulabandhula,et al. Incentivising Exploration and Recommendations for Contextual Bandits with Payments , 2020, EUMAS/AT.
[7] Aleksandrs Slivkins. Incentivizing exploration via information asymmetry , 2017, XRDS.
[8] Tor Lattimore,et al. The End of Optimism? An Asymptotic Analysis of Finite-Armed Linear Bandits , 2016, AISTATS.
[9] Babak Hassibi,et al. Stochastic Linear Bandits with Hidden Low Rank Structure , 2019, ArXiv.
[10] Peter Auer,et al. Finite-time Analysis of the Multiarmed Bandit Problem , 2002, Machine Learning.
[11] Aleksandrs Slivkins,et al. Sample Complexity of Incentivized Exploration , 2020, ArXiv.
[12] Jon M. Kleinberg,et al. Incentivizing exploration , 2014, EC.
[13] Sampath Kannan,et al. Fairness Incentives for Myopic Agents , 2017, EC.
[14] Andreas Krause,et al. Learning User Preferences to Incentivize Exploration in the Sharing Economy , 2017, AAAI.
[15] Shipra Agrawal,et al. Thompson Sampling for Contextual Bandits with Linear Payoffs , 2012, ICML.
[16] Bangrui Chen,et al. Incentivizing Exploration by Heterogeneous Users , 2018, COLT.
[17] Csaba Szepesvári,et al. Improved Algorithms for Linear Stochastic Bandits , 2011, NIPS.
[18] Peter Auer,et al. Using Confidence Bounds for Exploitation-Exploration Trade-offs , 2003, J. Mach. Learn. Res..
[19] Yishay Mansour,et al. Implementing the “Wisdom of the Crowd” , 2013, Journal of Political Economy.
[20] Simon S. Du,et al. Impact of Representation Learning in Linear Bandits , 2020, ICLR.
[21] Nicole Immorlica,et al. Incentivizing Exploration with Selective Data Disclosure , 2018, EC.
[22] Wei Chu,et al. Contextual Bandits with Linear Payoff Functions , 2011, AISTATS.
[23] Yishay Mansour,et al. Bayesian Incentive-Compatible Bandit Exploration , 2018 .
[24] Siwei Wang,et al. Multi-armed Bandits with Compensation , 2018, NeurIPS.