Journal articles on the topic 'Multiarmed Bandits'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 journal articles for your research on the topic 'Multiarmed Bandits.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse journal articles on a wide variety of disciplines and organise your bibliography correctly.
Righter, Rhonda, and J. George Shanthikumar. "Independently Expiring Multiarmed Bandits." Probability in the Engineering and Informational Sciences 12, no. 4 (October 1998): 453–68. http://dx.doi.org/10.1017/s0269964800005325.
Full textGao, Xiujuan, Hao Liang, and Tong Wang. "A Common Value Experimentation with Multiarmed Bandits." Mathematical Problems in Engineering 2018 (July 30, 2018): 1–8. http://dx.doi.org/10.1155/2018/4791590.
Full textKalathil, Dileep, Naumaan Nayyar, and Rahul Jain. "Decentralized Learning for Multiplayer Multiarmed Bandits." IEEE Transactions on Information Theory 60, no. 4 (April 2014): 2331–45. http://dx.doi.org/10.1109/tit.2014.2302471.
Full textCesa-Bianchi, Nicolò. "MULTIARMED BANDITS IN THE WORST CASE." IFAC Proceedings Volumes 35, no. 1 (2002): 91–96. http://dx.doi.org/10.3182/20020721-6-es-1901.01001.
Full textBray, Robert L., Decio Coviello, Andrea Ichino, and Nicola Persico. "Multitasking, Multiarmed Bandits, and the Italian Judiciary." Manufacturing & Service Operations Management 18, no. 4 (October 2016): 545–58. http://dx.doi.org/10.1287/msom.2016.0586.
Full textDenardo, Eric V., Haechurl Park, and Uriel G. Rothblum. "Risk-Sensitive and Risk-Neutral Multiarmed Bandits." Mathematics of Operations Research 32, no. 2 (May 2007): 374–94. http://dx.doi.org/10.1287/moor.1060.0240.
Full textWeber, Richard. "On the Gittins Index for Multiarmed Bandits." Annals of Applied Probability 2, no. 4 (November 1992): 1024–33. http://dx.doi.org/10.1214/aoap/1177005588.
Full textDrugan, Madalina M. "Covariance Matrix Adaptation for Multiobjective Multiarmed Bandits." IEEE Transactions on Neural Networks and Learning Systems 30, no. 8 (August 2019): 2493–502. http://dx.doi.org/10.1109/tnnls.2018.2885123.
Full textBurnetas, Apostolos N., and Michael N. Katehakis. "ASYMPTOTIC BAYES ANALYSIS FOR THE FINITE-HORIZON ONE-ARMED-BANDIT PROBLEM." Probability in the Engineering and Informational Sciences 17, no. 1 (January 2003): 53–82. http://dx.doi.org/10.1017/s0269964803171045.
Full textNayyar, Naumaan, Dileep Kalathil, and Rahul Jain. "On Regret-Optimal Learning in Decentralized Multiplayer Multiarmed Bandits." IEEE Transactions on Control of Network Systems 5, no. 1 (March 2018): 597–606. http://dx.doi.org/10.1109/tcns.2016.2635380.
Full textReverdy, Paul B., Vaibhav Srivastava, and Naomi Ehrich Leonard. "Modeling Human Decision Making in Generalized Gaussian Multiarmed Bandits." Proceedings of the IEEE 102, no. 4 (April 2014): 544–71. http://dx.doi.org/10.1109/jproc.2014.2307024.
Full textKrishnamurthy, Vikram, and Bo Wahlberg. "Partially Observed Markov Decision Process Multiarmed Bandits—Structural Results." Mathematics of Operations Research 34, no. 2 (May 2009): 287–302. http://dx.doi.org/10.1287/moor.1080.0371.
Full textCamerlenghi, Federico, Bianca Dumitrascu, Federico Ferrari, Barbara E. Engelhardt, and Stefano Favaro. "Nonparametric Bayesian multiarmed bandits for single-cell experiment design." Annals of Applied Statistics 14, no. 4 (December 2020): 2003–19. http://dx.doi.org/10.1214/20-aoas1370.
Full textMintz, Yonatan, Anil Aswani, Philip Kaminsky, Elena Flowers, and Yoshimi Fukuoka. "Nonstationary Bandits with Habituation and Recovery Dynamics." Operations Research 68, no. 5 (September 2020): 1493–516. http://dx.doi.org/10.1287/opre.2019.1918.
Full textGlazebrook, K. D., D. Ruiz-Hernandez, and C. Kirkbride. "Some indexable families of restless bandit problems." Advances in Applied Probability 38, no. 3 (September 2006): 643–72. http://dx.doi.org/10.1239/aap/1158684996.
Full textGlazebrook, K. D., D. Ruiz-Hernandez, and C. Kirkbride. "Some indexable families of restless bandit problems." Advances in Applied Probability 38, no. 03 (September 2006): 643–72. http://dx.doi.org/10.1017/s000186780000121x.
Full textMeshram, Rahul, D. Manjunath, and Aditya Gopalan. "On the Whittle Index for Restless Multiarmed Hidden Markov Bandits." IEEE Transactions on Automatic Control 63, no. 9 (September 2018): 3046–53. http://dx.doi.org/10.1109/tac.2018.2799521.
Full textCaro, Felipe, and Onesun Steve Yoo. "INDEXABILITY OF BANDIT PROBLEMS WITH RESPONSE DELAYS." Probability in the Engineering and Informational Sciences 24, no. 3 (April 23, 2010): 349–74. http://dx.doi.org/10.1017/s0269964810000021.
Full textGlazebrook, K. D., and R. Minty. "A Generalized Gittins Index for a Class of Multiarmed Bandits with General Resource Requirements." Mathematics of Operations Research 34, no. 1 (February 2009): 26–44. http://dx.doi.org/10.1287/moor.1080.0342.
Full textFarias, Vivek F., and Ritesh Madan. "The Irrevocable Multiarmed Bandit Problem." Operations Research 59, no. 2 (April 2011): 383–99. http://dx.doi.org/10.1287/opre.1100.0891.
Full textAuer, Peter, Nicolò Cesa-Bianchi, Yoav Freund, and Robert E. Schapire. "The Nonstochastic Multiarmed Bandit Problem." SIAM Journal on Computing 32, no. 1 (January 2002): 48–77. http://dx.doi.org/10.1137/s0097539701398375.
Full textPeköz, Erol A. "Some memoryless bandit policies." Journal of Applied Probability 40, no. 1 (March 2003): 250–56. http://dx.doi.org/10.1239/jap/1044476838.
Full textPeköz, Erol A. "Some memoryless bandit policies." Journal of Applied Probability 40, no. 01 (March 2003): 250–56. http://dx.doi.org/10.1017/s0021900200022373.
Full textDayanik, Savas, Warren Powell, and Kazutoshi Yamazaki. "Index policies for discounted bandit problems with availability constraints." Advances in Applied Probability 40, no. 2 (June 2008): 377–400. http://dx.doi.org/10.1239/aap/1214950209.
Full textDayanik, Savas, Warren Powell, and Kazutoshi Yamazaki. "Index policies for discounted bandit problems with availability constraints." Advances in Applied Probability 40, no. 02 (June 2008): 377–400. http://dx.doi.org/10.1017/s0001867800002573.
Full textTsitsiklis, J. "A lemma on the multiarmed bandit problem." IEEE Transactions on Automatic Control 31, no. 6 (June 1986): 576–77. http://dx.doi.org/10.1109/tac.1986.1104332.
Full textReverdy, Paul, Vaibhav Srivastava, and Naomi Ehrich Leonard. "Corrections to “Satisficing in Multiarmed Bandit Problems”." IEEE Transactions on Automatic Control 66, no. 1 (January 2021): 476–78. http://dx.doi.org/10.1109/tac.2020.2981433.
Full textFrostig, Esther, and Gideon Weiss. "Four proofs of Gittins’ multiarmed bandit theorem." Annals of Operations Research 241, no. 1-2 (January 7, 2014): 127–65. http://dx.doi.org/10.1007/s10479-013-1523-0.
Full textIshikida, Takashi, and Yat-wah Wan. "Scheduling Jobs That Are Subject to Deterministic Due Dates and Have Deteriorating Expected Rewards." Probability in the Engineering and Informational Sciences 11, no. 1 (January 1997): 65–78. http://dx.doi.org/10.1017/s026996480000468x.
Full textJiang, Weijin, Pingping Chen, Wanqing Zhang, Yongxia Sun, Chen Junpeng, and Qing Wen. "User Recruitment Algorithm for Maximizing Quality under Limited Budget in Mobile Crowdsensing." Discrete Dynamics in Nature and Society 2022 (January 20, 2022): 1–13. http://dx.doi.org/10.1155/2022/4804231.
Full textZeng, Fanzi, and Xinwang Shen. "Channel Selection Based on Trust and Multiarmed Bandit in Multiuser, Multichannel Cognitive Radio Networks." Scientific World Journal 2014 (2014): 1–6. http://dx.doi.org/10.1155/2014/916156.
Full textMersereau, A. J., P. Rusmevichientong, and J. N. Tsitsiklis. "A Structured Multiarmed Bandit Problem and the Greedy Policy." IEEE Transactions on Automatic Control 54, no. 12 (December 2009): 2787–802. http://dx.doi.org/10.1109/tac.2009.2031725.
Full textVaraiya, P., J. Walrand, and C. Buyukkoc. "Extensions of the multiarmed bandit problem: The discounted case." IEEE Transactions on Automatic Control 30, no. 5 (May 1985): 426–39. http://dx.doi.org/10.1109/tac.1985.1103989.
Full textMartin, David M., and Fred A. Johnson. "A Multiarmed Bandit Approach to Adaptive Water Quality Management." Integrated Environmental Assessment and Management 16, no. 6 (August 14, 2020): 841–52. http://dx.doi.org/10.1002/ieam.4302.
Full textKang, Xiaohan, Hong Ri, Mohd Nor Akmal Khalid, and Hiroyuki Iida. "Addictive Games: Case Study on Multi-Armed Bandit Game." Information 12, no. 12 (December 15, 2021): 521. http://dx.doi.org/10.3390/info12120521.
Full textMeng, Hao, Wasswa Shafik, S. Mojtaba Matinkhah, and Zubair Ahmad. "A 5G Beam Selection Machine Learning Algorithm for Unmanned Aerial Vehicle Applications." Wireless Communications and Mobile Computing 2020 (August 1, 2020): 1–16. http://dx.doi.org/10.1155/2020/1428968.
Full textChang, Hyeong Soo, and Sanghee Choe. "Combining Multiple Strategies for Multiarmed Bandit Problems and Asymptotic Optimality." Journal of Control Science and Engineering 2015 (2015): 1–7. http://dx.doi.org/10.1155/2015/264953.
Full textYoshida, Y. "Optimal stopping problems for multiarmed bandit processes with arms' independence." Computers & Mathematics with Applications 26, no. 12 (December 1993): 47–60. http://dx.doi.org/10.1016/0898-1221(93)90058-4.
Full textGokcesu, Kaan, and Suleyman Serdar Kozat. "An Online Minimax Optimal Algorithm for Adversarial Multiarmed Bandit Problem." IEEE Transactions on Neural Networks and Learning Systems 29, no. 11 (November 2018): 5565–80. http://dx.doi.org/10.1109/tnnls.2018.2806006.
Full textMisra, Kanishka, Eric M. Schwartz, and Jacob Abernethy. "Dynamic Online Pricing with Incomplete Information Using Multiarmed Bandit Experiments." Marketing Science 38, no. 2 (March 2019): 226–52. http://dx.doi.org/10.1287/mksc.2018.1129.
Full textToelch, Ulf, Matthew J. Bruce, Marius T. H. Meeus, and Simon M. Reader. "Humans copy rapidly increasing choices in a multiarmed bandit problem." Evolution and Human Behavior 31, no. 5 (September 2010): 326–33. http://dx.doi.org/10.1016/j.evolhumbehav.2010.03.002.
Full textMuqattash, Isa, and Jiaqiao Hu. "An ϵ-Greedy Multiarmed Bandit Approach to Markov Decision Processes." Stats 6, no. 1 (January 1, 2023): 99–112. http://dx.doi.org/10.3390/stats6010006.
Full textMansour, Yishay, Aleksandrs Slivkins, and Vasilis Syrgkanis. "Bayesian Incentive-Compatible Bandit Exploration." Operations Research 68, no. 4 (July 2020): 1132–61. http://dx.doi.org/10.1287/opre.2019.1949.
Full textUriarte, Alberto, and Santiago Ontañón. "Improving Monte Carlo Tree Search Policies in StarCraft via Probabilistic Models Learned from Replay Data." Proceedings of the AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment 12, no. 1 (June 25, 2021): 100–106. http://dx.doi.org/10.1609/aiide.v12i1.12852.
Full textQu, Yuben, Chao Dong, Dawei Niu, Hai Wang, and Chang Tian. "A Two-Dimensional Multiarmed Bandit Approach to Secondary Users with Network Coding in Cognitive Radio Networks." Mathematical Problems in Engineering 2015 (2015): 1–10. http://dx.doi.org/10.1155/2015/672837.
Full textBao, Wenqing, Xiaoqiang Cai, and Xianyi Wu. "A General Theory of MultiArmed Bandit Processes with Constrained Arm Switches." SIAM Journal on Control and Optimization 59, no. 6 (January 2021): 4666–88. http://dx.doi.org/10.1137/19m1282386.
Full textDrabik, Ewa. "On nearly selfoptimizing strategies for multiarmed bandit problems with controlled arms." Applicationes Mathematicae 23, no. 4 (1996): 449–73. http://dx.doi.org/10.4064/am-23-4-449-473.
Full textLiu, Haoyang, Keqin Liu, and Qing Zhao. "Learning in a Changing World: Restless Multiarmed Bandit With Unknown Dynamics." IEEE Transactions on Information Theory 59, no. 3 (March 2013): 1902–16. http://dx.doi.org/10.1109/tit.2012.2230215.
Full textAgrawal, Himanshu, and Krishna Asawa. "Decentralized Learning for Opportunistic Spectrum Access: Multiuser Restless Multiarmed Bandit Formulation." IEEE Systems Journal 14, no. 2 (June 2020): 2485–96. http://dx.doi.org/10.1109/jsyst.2019.2943361.
Full textNakayama, Kazuaki, Ryuzo Nakamura, Masato Hisakado, and Shintaro Mori. "Optimal learning dynamics of multiagent system in restless multiarmed bandit game." Physica A: Statistical Mechanics and its Applications 549 (July 2020): 124314. http://dx.doi.org/10.1016/j.physa.2020.124314.
Full text