Journal articles on the topic 'Bandit learning'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 journal articles for your research on the topic 'Bandit learning.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse journal articles on a wide variety of disciplines and organise your bibliography correctly.
Ciucanu, Radu, Pascal Lafourcade, Gael Marcadet, and Marta Soare. "SAMBA: A Generic Framework for Secure Federated Multi-Armed Bandits." Journal of Artificial Intelligence Research 73 (February 23, 2022): 737–65. http://dx.doi.org/10.1613/jair.1.13163.
Full textSharaf, Amr, and Hal Daumé III. "Meta-Learning Effective Exploration Strategies for Contextual Bandits." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 11 (May 18, 2021): 9541–48. http://dx.doi.org/10.1609/aaai.v35i11.17149.
Full textYang, Luting, Jianyi Yang, and Shaolei Ren. "Contextual Bandits with Delayed Feedback and Semi-supervised Learning (Student Abstract)." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 18 (May 18, 2021): 15943–44. http://dx.doi.org/10.1609/aaai.v35i18.17968.
Full textKapoor, Sayash, Kumar Kshitij Patel, and Purushottam Kar. "Corruption-tolerant bandit learning." Machine Learning 108, no. 4 (August 29, 2018): 687–715. http://dx.doi.org/10.1007/s10994-018-5758-5.
Full textCheung, Wang Chi, David Simchi-Levi, and Ruihao Zhu. "Hedging the Drift: Learning to Optimize Under Nonstationarity." Management Science 68, no. 3 (March 2022): 1696–713. http://dx.doi.org/10.1287/mnsc.2021.4024.
Full textDu, Yihan, Siwei Wang, and Longbo Huang. "A One-Size-Fits-All Solution to Conservative Bandit Problems." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 8 (May 18, 2021): 7254–61. http://dx.doi.org/10.1609/aaai.v35i8.16891.
Full textNarita, Yusuke, Shota Yasui, and Kohei Yata. "Efficient Counterfactual Learning from Bandit Feedback." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 4634–41. http://dx.doi.org/10.1609/aaai.v33i01.33014634.
Full textLupu, Andrei, Audrey Durand, and Doina Precup. "Leveraging Observations in Bandits: Between Risks and Benefits." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 6112–19. http://dx.doi.org/10.1609/aaai.v33i01.33016112.
Full textLopez, Romain, Inderjit S. Dhillon, and Michael I. Jordan. "Learning from eXtreme Bandit Feedback." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 10 (May 18, 2021): 8732–40. http://dx.doi.org/10.1609/aaai.v35i10.17058.
Full textCaro, Felipe, and Onesun Steve Yoo. "INDEXABILITY OF BANDIT PROBLEMS WITH RESPONSE DELAYS." Probability in the Engineering and Informational Sciences 24, no. 3 (April 23, 2010): 349–74. http://dx.doi.org/10.1017/s0269964810000021.
Full textZhu, Zhaowei, Jingxuan Zhu, Ji Liu, and Yang Liu. "Federated Bandit." Proceedings of the ACM on Measurement and Analysis of Computing Systems 5, no. 1 (February 18, 2021): 1–29. http://dx.doi.org/10.1145/3447380.
Full textVaratharajah, Yogatheesan, and Brent Berry. "A Contextual-Bandit-Based Approach for Informed Decision-Making in Clinical Trials." Life 12, no. 8 (August 21, 2022): 1277. http://dx.doi.org/10.3390/life12081277.
Full textAsanov, Igor. "Bandit cascade: A test of observational learning in the bandit problem." Journal of Economic Behavior & Organization 189 (September 2021): 150–71. http://dx.doi.org/10.1016/j.jebo.2021.06.006.
Full textDimakopoulou, Maria, Zhengyuan Zhou, Susan Athey, and Guido Imbens. "Balanced Linear Contextual Bandits." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 3445–53. http://dx.doi.org/10.1609/aaai.v33i01.33013445.
Full textYang, Jianyi, and Shaolei Ren. "Robust Bandit Learning with Imperfect Context." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 12 (May 18, 2021): 10594–602. http://dx.doi.org/10.1609/aaai.v35i12.17267.
Full textTruong, Quoc-Tuan, and Hady W. Lauw. "Variational learning from implicit bandit feedback." Machine Learning 110, no. 8 (July 9, 2021): 2085–105. http://dx.doi.org/10.1007/s10994-021-06028-0.
Full textTze-Leung Lai and S. Yakowitz. "Machine learning and nonparametric bandit theory." IEEE Transactions on Automatic Control 40, no. 7 (July 1995): 1199–209. http://dx.doi.org/10.1109/9.400491.
Full textTran, Alasdair, Cheng Soon Ong, and Christian Wolf. "Combining active learning suggestions." PeerJ Computer Science 4 (July 23, 2018): e157. http://dx.doi.org/10.7717/peerj-cs.157.
Full textShi, Chengshuai, and Cong Shen. "Federated Multi-Armed Bandits." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 11 (May 18, 2021): 9603–11. http://dx.doi.org/10.1609/aaai.v35i11.17156.
Full textNobari, Sadegh. "DBA: Dynamic Multi-Armed Bandit Algorithm." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 9869–70. http://dx.doi.org/10.1609/aaai.v33i01.33019869.
Full textKarpov, Nikolai, and Qin Zhang. "Instance-Sensitive Algorithms for Pure Exploration in Multinomial Logit Bandit." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 7 (June 28, 2022): 7096–103. http://dx.doi.org/10.1609/aaai.v36i7.20669.
Full textGao, Xuefeng, and Tianrun Xu. "Order scoring, bandit learning and order cancellations." Journal of Economic Dynamics and Control 134 (January 2022): 104287. http://dx.doi.org/10.1016/j.jedc.2021.104287.
Full textXu, Yiming, Vahid Keshavarzzadeh, Robert M. Kirby, and Akil Narayan. "A Bandit-Learning Approach to Multifidelity Approximation." SIAM Journal on Scientific Computing 44, no. 1 (January 18, 2022): A150—A175. http://dx.doi.org/10.1137/21m1408312.
Full textBrezzi, Monica, and Tze Leung Lai. "Optimal learning and experimentation in bandit problems." Journal of Economic Dynamics and Control 27, no. 1 (November 2002): 87–108. http://dx.doi.org/10.1016/s0165-1889(01)00028-8.
Full textRosenberg, Dinah, Eilon Solan, and Nicolas Vieille. "Social Learning in One-Arm Bandit Problems." Econometrica 75, no. 6 (November 2007): 1591–611. http://dx.doi.org/10.1111/j.1468-0262.2007.00807.x.
Full textLefebvre, Germain, Christopher Summerfield, and Rafal Bogacz. "A Normative Account of Confirmation Bias During Reinforcement Learning." Neural Computation 34, no. 2 (January 14, 2022): 307–37. http://dx.doi.org/10.1162/neco_a_01455.
Full textHuo, Xiaoguang, and Feng Fu. "Risk-aware multi-armed bandit problem with application to portfolio selection." Royal Society Open Science 4, no. 11 (November 2017): 171377. http://dx.doi.org/10.1098/rsos.171377.
Full textZhu, Zhaowei, Jingxuan Zhu, Ji Liu, and Yang Liu. "Federated Bandit: A Gossiping Approach." ACM SIGMETRICS Performance Evaluation Review 49, no. 1 (June 22, 2022): 3–4. http://dx.doi.org/10.1145/3543516.3453919.
Full textKaibel, Chris, and Torsten Biemann. "Rethinking the Gold Standard With Multi-armed Bandits: Machine Learning Allocation Algorithms for Experiments." Organizational Research Methods 24, no. 1 (June 11, 2019): 78–103. http://dx.doi.org/10.1177/1094428119854153.
Full textWu, Wen, Nan Cheng, Ning Zhang, Peng Yang, Weihua Zhuang, and Xuemin Shen. "Fast mmwave Beam Alignment via Correlated Bandit Learning." IEEE Transactions on Wireless Communications 18, no. 12 (December 2019): 5894–908. http://dx.doi.org/10.1109/twc.2019.2940454.
Full textHe, Di, Wei Chen, Liwei Wang, and Tie-Yan Liu. "Online learning for auction mechanism in bandit setting." Decision Support Systems 56 (December 2013): 379–86. http://dx.doi.org/10.1016/j.dss.2013.07.004.
Full textCayci, Semih, Atilla Eryilmaz, and R. Srikant. "Learning to Control Renewal Processes with Bandit Feedback." ACM SIGMETRICS Performance Evaluation Review 47, no. 1 (December 17, 2019): 41–42. http://dx.doi.org/10.1145/3376930.3376957.
Full textCayci, Semih, Atilla Eryilmaz, and R. Srikant. "Learning to Control Renewal Processes with Bandit Feedback." Proceedings of the ACM on Measurement and Analysis of Computing Systems 3, no. 2 (June 19, 2019): 1–32. http://dx.doi.org/10.1145/3341617.3326158.
Full textTang, Qiao, Hong Xie, Yunni Xia, Jia Lee, and Qingsheng Zhu. "Robust Contextual Bandits via Bootstrapping." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 13 (May 18, 2021): 12182–89. http://dx.doi.org/10.1609/aaai.v35i13.17446.
Full textGarcelon, Evrard, Mohammad Ghavamzadeh, Alessandro Lazaric, and Matteo Pirotta. "Improved Algorithms for Conservative Exploration in Bandits." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 04 (April 3, 2020): 3962–69. http://dx.doi.org/10.1609/aaai.v34i04.5812.
Full textGiachino, Chiara, Luigi Bollani, Alessandro Bonadonna, and Marco Bertetti. "Reinforcement learning for content's customization: a first step of experimentation in Skyscanner." Industrial Management & Data Systems 121, no. 6 (January 15, 2021): 1417–34. http://dx.doi.org/10.1108/imds-12-2019-0722.
Full textMetelli, Alberto Maria, Matteo Papini, Pierluca D'Oro, and Marcello Restelli. "Policy Optimization as Online Learning with Mediator Feedback." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 10 (May 18, 2021): 8958–66. http://dx.doi.org/10.1609/aaai.v35i10.17083.
Full textShi, Zheyuan Ryan, Zhiwei Steven Wu, Rayid Ghani, and Fei Fang. "Bandit Data-Driven Optimization for Crowdsourcing Food Rescue Platforms." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 11 (June 28, 2022): 12154–62. http://dx.doi.org/10.1609/aaai.v36i11.21475.
Full textDing, Qinxu, Yong Liu, Chunyan Miao, Fei Cheng, and Haihong Tang. "A Hybrid Bandit Framework for Diversified Recommendation." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 5 (May 18, 2021): 4036–44. http://dx.doi.org/10.1609/aaai.v35i5.16524.
Full textModaresi, Sajad, Denis Sauré, and Juan Pablo Vielma. "Learning in Combinatorial Optimization: What and How to Explore." Operations Research 68, no. 5 (September 2020): 1585–604. http://dx.doi.org/10.1287/opre.2019.1926.
Full textYan, Cairong, Junli Xian, Yongquan Wan, and Pengwei Wang. "Modeling implicit feedback based on bandit learning for recommendation." Neurocomputing 447 (August 2021): 244–56. http://dx.doi.org/10.1016/j.neucom.2021.03.072.
Full textChen, Lixing, Jie Xu, Shaolei Ren, and Pan Zhou. "Spatio–Temporal Edge Service Placement: A Bandit Learning Approach." IEEE Transactions on Wireless Communications 17, no. 12 (December 2018): 8388–401. http://dx.doi.org/10.1109/twc.2018.2876823.
Full textXia, Wenchao, Tony Q. S. Quek, Kun Guo, Wanli Wen, Howard H. Yang, and Hongbo Zhu. "Multi-Armed Bandit-Based Client Scheduling for Federated Learning." IEEE Transactions on Wireless Communications 19, no. 11 (November 2020): 7108–23. http://dx.doi.org/10.1109/twc.2020.3008091.
Full textKaufmann, Emilie, and Aurélien Garivier. "Learning the distribution with largest mean: two bandit frameworks." ESAIM: Proceedings and Surveys 60 (2017): 114–31. http://dx.doi.org/10.1051/proc/201760114.
Full textAgrawal, Shipra, Vashist Avadhanula, Vineet Goyal, and Assaf Zeevi. "MNL-Bandit: A Dynamic Learning Approach to Assortment Selection." Operations Research 67, no. 5 (September 2019): 1453–85. http://dx.doi.org/10.1287/opre.2018.1832.
Full textLiu, Keqin, and Qing Zhao. "Distributed Learning in Multi-Armed Bandit With Multiple Players." IEEE Transactions on Signal Processing 58, no. 11 (November 2010): 5667–81. http://dx.doi.org/10.1109/tsp.2010.2062509.
Full textHan, Kai, Yuntian He, Alex X. Liu, Shaojie Tang, and He Huang. "Differentially Private and Budget-Limited Bandit Learning over Matroids." INFORMS Journal on Computing 32, no. 3 (July 2020): 790–804. http://dx.doi.org/10.1287/ijoc.2019.0903.
Full textJain, Shweta, Satyanath Bhat, Ganesh Ghalme, Divya Padmanabhan, and Y. Narahari. "Mechanisms with learning for stochastic multi-armed bandit problems." Indian Journal of Pure and Applied Mathematics 47, no. 2 (June 2016): 229–72. http://dx.doi.org/10.1007/s13226-016-0186-3.
Full textMeng, Hao, Wasswa Shafik, S. Mojtaba Matinkhah, and Zubair Ahmad. "A 5G Beam Selection Machine Learning Algorithm for Unmanned Aerial Vehicle Applications." Wireless Communications and Mobile Computing 2020 (August 1, 2020): 1–16. http://dx.doi.org/10.1155/2020/1428968.
Full textYan, Xue, Yali Du, Binxin Ru, Jun Wang, Haifeng Zhang, and Xu Chen. "Learning to Identify Top Elo Ratings: A Dueling Bandits Approach." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 8 (June 28, 2022): 8797–805. http://dx.doi.org/10.1609/aaai.v36i8.20860.
Full text