Статті в журналах з теми "Bandit learning"
Оформте джерело за APA, MLA, Chicago, Harvard та іншими стилями
Ознайомтеся з топ-50 статей у журналах для дослідження на тему "Bandit learning".
Біля кожної праці в переліку літератури доступна кнопка «Додати до бібліографії». Скористайтеся нею – і ми автоматично оформимо бібліографічне посилання на обрану працю в потрібному вам стилі цитування: APA, MLA, «Гарвард», «Чикаго», «Ванкувер» тощо.
Також ви можете завантажити повний текст наукової публікації у форматі «.pdf» та прочитати онлайн анотацію до роботи, якщо відповідні параметри наявні в метаданих.
Переглядайте статті в журналах для різних дисциплін та оформлюйте правильно вашу бібліографію.
Ciucanu, Radu, Pascal Lafourcade, Gael Marcadet, and Marta Soare. "SAMBA: A Generic Framework for Secure Federated Multi-Armed Bandits." Journal of Artificial Intelligence Research 73 (February 23, 2022): 737–65. http://dx.doi.org/10.1613/jair.1.13163.
Повний текст джерелаSharaf, Amr, and Hal Daumé III. "Meta-Learning Effective Exploration Strategies for Contextual Bandits." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 11 (May 18, 2021): 9541–48. http://dx.doi.org/10.1609/aaai.v35i11.17149.
Повний текст джерелаYang, Luting, Jianyi Yang, and Shaolei Ren. "Contextual Bandits with Delayed Feedback and Semi-supervised Learning (Student Abstract)." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 18 (May 18, 2021): 15943–44. http://dx.doi.org/10.1609/aaai.v35i18.17968.
Повний текст джерелаKapoor, Sayash, Kumar Kshitij Patel, and Purushottam Kar. "Corruption-tolerant bandit learning." Machine Learning 108, no. 4 (August 29, 2018): 687–715. http://dx.doi.org/10.1007/s10994-018-5758-5.
Повний текст джерелаCheung, Wang Chi, David Simchi-Levi, and Ruihao Zhu. "Hedging the Drift: Learning to Optimize Under Nonstationarity." Management Science 68, no. 3 (March 2022): 1696–713. http://dx.doi.org/10.1287/mnsc.2021.4024.
Повний текст джерелаDu, Yihan, Siwei Wang, and Longbo Huang. "A One-Size-Fits-All Solution to Conservative Bandit Problems." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 8 (May 18, 2021): 7254–61. http://dx.doi.org/10.1609/aaai.v35i8.16891.
Повний текст джерелаNarita, Yusuke, Shota Yasui, and Kohei Yata. "Efficient Counterfactual Learning from Bandit Feedback." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 4634–41. http://dx.doi.org/10.1609/aaai.v33i01.33014634.
Повний текст джерелаLupu, Andrei, Audrey Durand, and Doina Precup. "Leveraging Observations in Bandits: Between Risks and Benefits." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 6112–19. http://dx.doi.org/10.1609/aaai.v33i01.33016112.
Повний текст джерелаLopez, Romain, Inderjit S. Dhillon, and Michael I. Jordan. "Learning from eXtreme Bandit Feedback." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 10 (May 18, 2021): 8732–40. http://dx.doi.org/10.1609/aaai.v35i10.17058.
Повний текст джерелаCaro, Felipe, and Onesun Steve Yoo. "INDEXABILITY OF BANDIT PROBLEMS WITH RESPONSE DELAYS." Probability in the Engineering and Informational Sciences 24, no. 3 (April 23, 2010): 349–74. http://dx.doi.org/10.1017/s0269964810000021.
Повний текст джерелаZhu, Zhaowei, Jingxuan Zhu, Ji Liu, and Yang Liu. "Federated Bandit." Proceedings of the ACM on Measurement and Analysis of Computing Systems 5, no. 1 (February 18, 2021): 1–29. http://dx.doi.org/10.1145/3447380.
Повний текст джерелаVaratharajah, Yogatheesan, and Brent Berry. "A Contextual-Bandit-Based Approach for Informed Decision-Making in Clinical Trials." Life 12, no. 8 (August 21, 2022): 1277. http://dx.doi.org/10.3390/life12081277.
Повний текст джерелаAsanov, Igor. "Bandit cascade: A test of observational learning in the bandit problem." Journal of Economic Behavior & Organization 189 (September 2021): 150–71. http://dx.doi.org/10.1016/j.jebo.2021.06.006.
Повний текст джерелаDimakopoulou, Maria, Zhengyuan Zhou, Susan Athey, and Guido Imbens. "Balanced Linear Contextual Bandits." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 3445–53. http://dx.doi.org/10.1609/aaai.v33i01.33013445.
Повний текст джерелаYang, Jianyi, and Shaolei Ren. "Robust Bandit Learning with Imperfect Context." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 12 (May 18, 2021): 10594–602. http://dx.doi.org/10.1609/aaai.v35i12.17267.
Повний текст джерелаTruong, Quoc-Tuan, and Hady W. Lauw. "Variational learning from implicit bandit feedback." Machine Learning 110, no. 8 (July 9, 2021): 2085–105. http://dx.doi.org/10.1007/s10994-021-06028-0.
Повний текст джерелаTze-Leung Lai and S. Yakowitz. "Machine learning and nonparametric bandit theory." IEEE Transactions on Automatic Control 40, no. 7 (July 1995): 1199–209. http://dx.doi.org/10.1109/9.400491.
Повний текст джерелаTran, Alasdair, Cheng Soon Ong, and Christian Wolf. "Combining active learning suggestions." PeerJ Computer Science 4 (July 23, 2018): e157. http://dx.doi.org/10.7717/peerj-cs.157.
Повний текст джерелаShi, Chengshuai, and Cong Shen. "Federated Multi-Armed Bandits." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 11 (May 18, 2021): 9603–11. http://dx.doi.org/10.1609/aaai.v35i11.17156.
Повний текст джерелаNobari, Sadegh. "DBA: Dynamic Multi-Armed Bandit Algorithm." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 9869–70. http://dx.doi.org/10.1609/aaai.v33i01.33019869.
Повний текст джерелаKarpov, Nikolai, and Qin Zhang. "Instance-Sensitive Algorithms for Pure Exploration in Multinomial Logit Bandit." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 7 (June 28, 2022): 7096–103. http://dx.doi.org/10.1609/aaai.v36i7.20669.
Повний текст джерелаGao, Xuefeng, and Tianrun Xu. "Order scoring, bandit learning and order cancellations." Journal of Economic Dynamics and Control 134 (January 2022): 104287. http://dx.doi.org/10.1016/j.jedc.2021.104287.
Повний текст джерелаXu, Yiming, Vahid Keshavarzzadeh, Robert M. Kirby, and Akil Narayan. "A Bandit-Learning Approach to Multifidelity Approximation." SIAM Journal on Scientific Computing 44, no. 1 (January 18, 2022): A150—A175. http://dx.doi.org/10.1137/21m1408312.
Повний текст джерелаBrezzi, Monica, and Tze Leung Lai. "Optimal learning and experimentation in bandit problems." Journal of Economic Dynamics and Control 27, no. 1 (November 2002): 87–108. http://dx.doi.org/10.1016/s0165-1889(01)00028-8.
Повний текст джерелаRosenberg, Dinah, Eilon Solan, and Nicolas Vieille. "Social Learning in One-Arm Bandit Problems." Econometrica 75, no. 6 (November 2007): 1591–611. http://dx.doi.org/10.1111/j.1468-0262.2007.00807.x.
Повний текст джерелаLefebvre, Germain, Christopher Summerfield, and Rafal Bogacz. "A Normative Account of Confirmation Bias During Reinforcement Learning." Neural Computation 34, no. 2 (January 14, 2022): 307–37. http://dx.doi.org/10.1162/neco_a_01455.
Повний текст джерелаHuo, Xiaoguang, and Feng Fu. "Risk-aware multi-armed bandit problem with application to portfolio selection." Royal Society Open Science 4, no. 11 (November 2017): 171377. http://dx.doi.org/10.1098/rsos.171377.
Повний текст джерелаZhu, Zhaowei, Jingxuan Zhu, Ji Liu, and Yang Liu. "Federated Bandit: A Gossiping Approach." ACM SIGMETRICS Performance Evaluation Review 49, no. 1 (June 22, 2022): 3–4. http://dx.doi.org/10.1145/3543516.3453919.
Повний текст джерелаKaibel, Chris, and Torsten Biemann. "Rethinking the Gold Standard With Multi-armed Bandits: Machine Learning Allocation Algorithms for Experiments." Organizational Research Methods 24, no. 1 (June 11, 2019): 78–103. http://dx.doi.org/10.1177/1094428119854153.
Повний текст джерелаWu, Wen, Nan Cheng, Ning Zhang, Peng Yang, Weihua Zhuang, and Xuemin Shen. "Fast mmwave Beam Alignment via Correlated Bandit Learning." IEEE Transactions on Wireless Communications 18, no. 12 (December 2019): 5894–908. http://dx.doi.org/10.1109/twc.2019.2940454.
Повний текст джерелаHe, Di, Wei Chen, Liwei Wang, and Tie-Yan Liu. "Online learning for auction mechanism in bandit setting." Decision Support Systems 56 (December 2013): 379–86. http://dx.doi.org/10.1016/j.dss.2013.07.004.
Повний текст джерелаCayci, Semih, Atilla Eryilmaz, and R. Srikant. "Learning to Control Renewal Processes with Bandit Feedback." ACM SIGMETRICS Performance Evaluation Review 47, no. 1 (December 17, 2019): 41–42. http://dx.doi.org/10.1145/3376930.3376957.
Повний текст джерелаCayci, Semih, Atilla Eryilmaz, and R. Srikant. "Learning to Control Renewal Processes with Bandit Feedback." Proceedings of the ACM on Measurement and Analysis of Computing Systems 3, no. 2 (June 19, 2019): 1–32. http://dx.doi.org/10.1145/3341617.3326158.
Повний текст джерелаTang, Qiao, Hong Xie, Yunni Xia, Jia Lee, and Qingsheng Zhu. "Robust Contextual Bandits via Bootstrapping." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 13 (May 18, 2021): 12182–89. http://dx.doi.org/10.1609/aaai.v35i13.17446.
Повний текст джерелаGarcelon, Evrard, Mohammad Ghavamzadeh, Alessandro Lazaric, and Matteo Pirotta. "Improved Algorithms for Conservative Exploration in Bandits." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 04 (April 3, 2020): 3962–69. http://dx.doi.org/10.1609/aaai.v34i04.5812.
Повний текст джерелаGiachino, Chiara, Luigi Bollani, Alessandro Bonadonna, and Marco Bertetti. "Reinforcement learning for content's customization: a first step of experimentation in Skyscanner." Industrial Management & Data Systems 121, no. 6 (January 15, 2021): 1417–34. http://dx.doi.org/10.1108/imds-12-2019-0722.
Повний текст джерелаMetelli, Alberto Maria, Matteo Papini, Pierluca D'Oro, and Marcello Restelli. "Policy Optimization as Online Learning with Mediator Feedback." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 10 (May 18, 2021): 8958–66. http://dx.doi.org/10.1609/aaai.v35i10.17083.
Повний текст джерелаShi, Zheyuan Ryan, Zhiwei Steven Wu, Rayid Ghani, and Fei Fang. "Bandit Data-Driven Optimization for Crowdsourcing Food Rescue Platforms." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 11 (June 28, 2022): 12154–62. http://dx.doi.org/10.1609/aaai.v36i11.21475.
Повний текст джерелаDing, Qinxu, Yong Liu, Chunyan Miao, Fei Cheng, and Haihong Tang. "A Hybrid Bandit Framework for Diversified Recommendation." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 5 (May 18, 2021): 4036–44. http://dx.doi.org/10.1609/aaai.v35i5.16524.
Повний текст джерелаModaresi, Sajad, Denis Sauré, and Juan Pablo Vielma. "Learning in Combinatorial Optimization: What and How to Explore." Operations Research 68, no. 5 (September 2020): 1585–604. http://dx.doi.org/10.1287/opre.2019.1926.
Повний текст джерелаYan, Cairong, Junli Xian, Yongquan Wan, and Pengwei Wang. "Modeling implicit feedback based on bandit learning for recommendation." Neurocomputing 447 (August 2021): 244–56. http://dx.doi.org/10.1016/j.neucom.2021.03.072.
Повний текст джерелаChen, Lixing, Jie Xu, Shaolei Ren, and Pan Zhou. "Spatio–Temporal Edge Service Placement: A Bandit Learning Approach." IEEE Transactions on Wireless Communications 17, no. 12 (December 2018): 8388–401. http://dx.doi.org/10.1109/twc.2018.2876823.
Повний текст джерелаXia, Wenchao, Tony Q. S. Quek, Kun Guo, Wanli Wen, Howard H. Yang, and Hongbo Zhu. "Multi-Armed Bandit-Based Client Scheduling for Federated Learning." IEEE Transactions on Wireless Communications 19, no. 11 (November 2020): 7108–23. http://dx.doi.org/10.1109/twc.2020.3008091.
Повний текст джерелаKaufmann, Emilie, and Aurélien Garivier. "Learning the distribution with largest mean: two bandit frameworks." ESAIM: Proceedings and Surveys 60 (2017): 114–31. http://dx.doi.org/10.1051/proc/201760114.
Повний текст джерелаAgrawal, Shipra, Vashist Avadhanula, Vineet Goyal, and Assaf Zeevi. "MNL-Bandit: A Dynamic Learning Approach to Assortment Selection." Operations Research 67, no. 5 (September 2019): 1453–85. http://dx.doi.org/10.1287/opre.2018.1832.
Повний текст джерелаLiu, Keqin, and Qing Zhao. "Distributed Learning in Multi-Armed Bandit With Multiple Players." IEEE Transactions on Signal Processing 58, no. 11 (November 2010): 5667–81. http://dx.doi.org/10.1109/tsp.2010.2062509.
Повний текст джерелаHan, Kai, Yuntian He, Alex X. Liu, Shaojie Tang, and He Huang. "Differentially Private and Budget-Limited Bandit Learning over Matroids." INFORMS Journal on Computing 32, no. 3 (July 2020): 790–804. http://dx.doi.org/10.1287/ijoc.2019.0903.
Повний текст джерелаJain, Shweta, Satyanath Bhat, Ganesh Ghalme, Divya Padmanabhan, and Y. Narahari. "Mechanisms with learning for stochastic multi-armed bandit problems." Indian Journal of Pure and Applied Mathematics 47, no. 2 (June 2016): 229–72. http://dx.doi.org/10.1007/s13226-016-0186-3.
Повний текст джерелаMeng, Hao, Wasswa Shafik, S. Mojtaba Matinkhah, and Zubair Ahmad. "A 5G Beam Selection Machine Learning Algorithm for Unmanned Aerial Vehicle Applications." Wireless Communications and Mobile Computing 2020 (August 1, 2020): 1–16. http://dx.doi.org/10.1155/2020/1428968.
Повний текст джерелаYan, Xue, Yali Du, Binxin Ru, Jun Wang, Haifeng Zhang, and Xu Chen. "Learning to Identify Top Elo Ratings: A Dueling Bandits Approach." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 8 (June 28, 2022): 8797–805. http://dx.doi.org/10.1609/aaai.v36i8.20860.
Повний текст джерела