Статті в журналах з теми "RL ALGORITHMS"
Оформте джерело за APA, MLA, Chicago, Harvard та іншими стилями
Ознайомтеся з топ-50 статей у журналах для дослідження на тему "RL ALGORITHMS".
Біля кожної праці в переліку літератури доступна кнопка «Додати до бібліографії». Скористайтеся нею – і ми автоматично оформимо бібліографічне посилання на обрану працю в потрібному вам стилі цитування: APA, MLA, «Гарвард», «Чикаго», «Ванкувер» тощо.
Також ви можете завантажити повний текст наукової публікації у форматі «.pdf» та прочитати онлайн анотацію до роботи, якщо відповідні параметри наявні в метаданих.
Переглядайте статті в журналах для різних дисциплін та оформлюйте правильно вашу бібліографію.
Lahande, Prathamesh, Parag Kaveri, and Jatinderkumar Saini. "Reinforcement Learning for Reducing the Interruptions and Increasing Fault Tolerance in the Cloud Environment." Informatics 10, no. 3 (August 2, 2023): 64. http://dx.doi.org/10.3390/informatics10030064.
Повний текст джерелаTrella, Anna L., Kelly W. Zhang, Inbal Nahum-Shani, Vivek Shetty, Finale Doshi-Velez, and Susan A. Murphy. "Designing Reinforcement Learning Algorithms for Digital Interventions: Pre-Implementation Guidelines." Algorithms 15, no. 8 (July 22, 2022): 255. http://dx.doi.org/10.3390/a15080255.
Повний текст джерелаRodríguez Sánchez, Francisco, Ildeberto Santos-Ruiz, Joaquín Domínguez-Zenteno, and Francisco Ronay López-Estrada. "Control Applications Using Reinforcement Learning: An Overview." Memorias del Congreso Nacional de Control Automático 5, no. 1 (October 17, 2022): 67–72. http://dx.doi.org/10.58571/cnca.amca.2022.019.
Повний текст джерелаAbbass, Mahmoud Abdelkader Bashery, and Hyun-Soo Kang. "Drone Elevation Control Based on Python-Unity Integrated Framework for Reinforcement Learning Applications." Drones 7, no. 4 (March 24, 2023): 225. http://dx.doi.org/10.3390/drones7040225.
Повний текст джерелаMann, Timothy, and Yoonsuck Choe. "Scaling Up Reinforcement Learning through Targeted Exploration." Proceedings of the AAAI Conference on Artificial Intelligence 25, no. 1 (August 4, 2011): 435–40. http://dx.doi.org/10.1609/aaai.v25i1.7929.
Повний текст джерелаCheng, Richard, Gábor Orosz, Richard M. Murray, and Joel W. Burdick. "End-to-End Safe Reinforcement Learning through Barrier Functions for Safety-Critical Continuous Control Tasks." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 3387–95. http://dx.doi.org/10.1609/aaai.v33i01.33013387.
Повний текст джерелаKirsch, Louis, Sebastian Flennerhag, Hado van Hasselt, Abram Friesen, Junhyuk Oh, and Yutian Chen. "Introducing Symmetries to Black Box Meta Reinforcement Learning." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 7 (June 28, 2022): 7202–10. http://dx.doi.org/10.1609/aaai.v36i7.20681.
Повний текст джерелаKim, Hyun-Su, and Uksun Kim. "Development of a Control Algorithm for a Semi-Active Mid-Story Isolation System Using Reinforcement Learning." Applied Sciences 13, no. 4 (February 4, 2023): 2053. http://dx.doi.org/10.3390/app13042053.
Повний текст джерелаPrakash, Kritika, Fiza Husain, Praveen Paruchuri, and Sujit Gujar. "How Private Is Your RL Policy? An Inverse RL Based Analysis Framework." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 7 (June 28, 2022): 8009–16. http://dx.doi.org/10.1609/aaai.v36i7.20772.
Повний текст джерелаNiazi, Abdolkarim, Norizah Redzuan, Raja Ishak Raja Hamzah, and Sara Esfandiari. "Improvement on Supporting Machine Learning Algorithm for Solving Problem in Immediate Decision Making." Advanced Materials Research 566 (September 2012): 572–79. http://dx.doi.org/10.4028/www.scientific.net/amr.566.572.
Повний текст джерелаMu, Tong, Georgios Theocharous, David Arbour, and Emma Brunskill. "Constraint Sampling Reinforcement Learning: Incorporating Expertise for Faster Learning." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 7 (June 28, 2022): 7841–49. http://dx.doi.org/10.1609/aaai.v36i7.20753.
Повний текст джерелаKołota, Jakub, and Turhan Can Kargin. "Comparison of Various Reinforcement Learning Environments in the Context of Continuum Robot Control." Applied Sciences 13, no. 16 (August 11, 2023): 9153. http://dx.doi.org/10.3390/app13169153.
Повний текст джерелаJang, Sun-Ho, Woo-Jin Ahn, Yu-Jin Kim, Hyung-Gil Hong, Dong-Sung Pae, and Myo-Taeg Lim. "Stable and Efficient Reinforcement Learning Method for Avoidance Driving of Unmanned Vehicles." Electronics 12, no. 18 (September 6, 2023): 3773. http://dx.doi.org/10.3390/electronics12183773.
Повний текст джерелаPeng, Zhiyong, Changlin Han, Yadong Liu, and Zongtan Zhou. "Weighted Policy Constraints for Offline Reinforcement Learning." Proceedings of the AAAI Conference on Artificial Intelligence 37, no. 8 (June 26, 2023): 9435–43. http://dx.doi.org/10.1609/aaai.v37i8.26130.
Повний текст джерелаTessler, Chen, Yuval Shpigelman, Gal Dalal, Amit Mandelbaum, Doron Haritan Kazakov, Benjamin Fuhrer, Gal Chechik, and Shie Mannor. "Reinforcement Learning for Datacenter Congestion Control." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 11 (June 28, 2022): 12615–21. http://dx.doi.org/10.1609/aaai.v36i11.21535.
Повний текст джерелаJIANG, JU, MOHAMED S. KAMEL, and LEI CHEN. "AGGREGATION OF MULTIPLE REINFORCEMENT LEARNING ALGORITHMS." International Journal on Artificial Intelligence Tools 15, no. 05 (October 2006): 855–61. http://dx.doi.org/10.1142/s0218213006002990.
Повний текст джерелаChen, Feng, Chenghe Wang, Fuxiang Zhang, Hao Ding, Qiaoyong Zhong, Shiliang Pu, and Zongzhang Zhang. "Towards Deployment-Efficient and Collision-Free Multi-Agent Path Finding (Student Abstract)." Proceedings of the AAAI Conference on Artificial Intelligence 37, no. 13 (June 26, 2023): 16182–83. http://dx.doi.org/10.1609/aaai.v37i13.26951.
Повний текст джерелаGuo, Kun, and Qishan Zhang. "A Discrete Artificial Bee Colony Algorithm for the Reverse Logistics Location and Routing Problem." International Journal of Information Technology & Decision Making 16, no. 05 (September 2017): 1339–57. http://dx.doi.org/10.1142/s0219622014500126.
Повний текст джерелаPadakandla, Sindhu. "A Survey of Reinforcement Learning Algorithms for Dynamically Varying Environments." ACM Computing Surveys 54, no. 6 (July 2021): 1–25. http://dx.doi.org/10.1145/3459991.
Повний текст джерелаGaon, Maor, and Ronen Brafman. "Reinforcement Learning with Non-Markovian Rewards." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 04 (April 3, 2020): 3980–87. http://dx.doi.org/10.1609/aaai.v34i04.5814.
Повний текст джерелаSun, Peiquan, Wengang Zhou, and Houqiang Li. "Attentive Experience Replay." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 04 (April 3, 2020): 5900–5907. http://dx.doi.org/10.1609/aaai.v34i04.6049.
Повний текст джерелаChen, Zaiwei. "A Unified Lyapunov Framework for Finite-Sample Analysis of Reinforcement Learning Algorithms." ACM SIGMETRICS Performance Evaluation Review 50, no. 3 (December 30, 2022): 12–15. http://dx.doi.org/10.1145/3579342.3579346.
Повний текст джерелаYau, Kok-Lim Alvin, Geong-Sen Poh, Su Fong Chien, and Hasan A. A. Al-Rawi. "Application of Reinforcement Learning in Cognitive Radio Networks: Models and Algorithms." Scientific World Journal 2014 (2014): 1–23. http://dx.doi.org/10.1155/2014/209810.
Повний текст джерелаTessler, Chen, Yuval Shpigelman, Gal Dalal, Amit Mandelbaum, Doron Haritan Kazakov, Benjamin Fuhrer, Gal Chechik, and Shie Mannor. "Reinforcement Learning for Datacenter Congestion Control." ACM SIGMETRICS Performance Evaluation Review 49, no. 2 (January 17, 2022): 43–46. http://dx.doi.org/10.1145/3512798.3512815.
Повний текст джерелаJin, Zengwang, Menglu Ma, Shuting Zhang, Yanyan Hu, Yanning Zhang, and Changyin Sun. "Secure State Estimation of Cyber-Physical System under Cyber Attacks: Q-Learning vs. SARSA." Electronics 11, no. 19 (October 1, 2022): 3161. http://dx.doi.org/10.3390/electronics11193161.
Повний текст джерелаLi, Shaodong, Xiaogang Yuan, and Jie Niu. "Robotic Peg-in-Hole Assembly Strategy Research Based on Reinforcement Learning Algorithm." Applied Sciences 12, no. 21 (November 3, 2022): 11149. http://dx.doi.org/10.3390/app122111149.
Повний текст джерелаPan, Yaozong, Jian Zhang, Chunhui Yuan, and Haitao Yang. "Supervised Reinforcement Learning via Value Function." Symmetry 11, no. 4 (April 24, 2019): 590. http://dx.doi.org/10.3390/sym11040590.
Повний текст джерелаKabanda, Professor Gabriel, Colletor Tendeukai Chipfumbu, and Tinashe Chingoriwo. "A Reinforcement Learning Paradigm for Cybersecurity Education and Training." Oriental journal of computer science and technology 16, no. 01 (May 30, 2023): 12–45. http://dx.doi.org/10.13005/ojcst16.01.02.
Повний текст джерелаYousif, Ayman Basheer, Hassan Jaleel Hassan, and Gaida Muttasher. "Applying reinforcement learning for random early detaction algorithm in adaptive queue management systems." Indonesian Journal of Electrical Engineering and Computer Science 26, no. 3 (June 1, 2022): 1684. http://dx.doi.org/10.11591/ijeecs.v26.i3.pp1684-1691.
Повний текст джерелаSzita, István, and András Lörincz. "Learning Tetris Using the Noisy Cross-Entropy Method." Neural Computation 18, no. 12 (December 2006): 2936–41. http://dx.doi.org/10.1162/neco.2006.18.12.2936.
Повний текст джерелаYe, Weicheng, and Dangxing Chen. "Analysis of Performance Measure in Q Learning with UCB Exploration." Mathematics 10, no. 4 (February 12, 2022): 575. http://dx.doi.org/10.3390/math10040575.
Повний текст джерелаLin, Xingbin, Deyu Yuan, and Xifei Li. "Reinforcement Learning with Dual Safety Policies for Energy Savings in Building Energy Systems." Buildings 13, no. 3 (February 21, 2023): 580. http://dx.doi.org/10.3390/buildings13030580.
Повний текст джерелаLi, Luchen, and A. Aldo Faisal. "Bayesian Distributional Policy Gradients." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 10 (May 18, 2021): 8429–37. http://dx.doi.org/10.1609/aaai.v35i10.17024.
Повний текст джерелаGrewal, Yashvir S., Frits De Nijs, and Sarah Goodwin. "Evaluating Meta-Reinforcement Learning through a HVAC Control Benchmark (Student Abstract)." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 18 (May 18, 2021): 15785–86. http://dx.doi.org/10.1609/aaai.v35i18.17889.
Повний текст джерелаVillalpando-Hernandez, Rafaela, Cesar Vargas-Rosales, and David Munoz-Rodriguez. "Localization Algorithm for 3D Sensor Networks: A Recursive Data Fusion Approach." Sensors 21, no. 22 (November 17, 2021): 7626. http://dx.doi.org/10.3390/s21227626.
Повний текст джерелаZhao, Richard, and Duane Szafron. "Learning Character Behaviors Using Agent Modeling in Games." Proceedings of the AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment 5, no. 1 (October 16, 2009): 179–85. http://dx.doi.org/10.1609/aiide.v5i1.12369.
Повний текст джерелаHu and Xu. "Fuzzy Reinforcement Learning and Curriculum Transfer Learning for Micromanagement in Multi-Robot Confrontation." Information 10, no. 11 (November 2, 2019): 341. http://dx.doi.org/10.3390/info10110341.
Повний текст джерелаShen, Haocheng, Jason Yosinski, Petar Kormushev, Darwin G. Caldwell, and Hod Lipson. "Learning Fast Quadruped Robot Gaits with the RL PoWER Spline Parameterization." Cybernetics and Information Technologies 12, no. 3 (September 1, 2012): 66–75. http://dx.doi.org/10.2478/cait-2012-0022.
Повний текст джерелаShaposhnikova, Sofiia, and Dmytro Omelian. "TOWARDS EFFECTIVE STRATEGIES FOR MOBILE ROBOT USING REINFORCEMENT LEARNING AND GRAPH ALGORITHMS." Automation of technological and business processes 15, no. 2 (June 19, 2023): 24–34. http://dx.doi.org/10.15673/atbp.v15i2.2522.
Повний текст джерелаLiao, Hanlin. "Urban Intersection Simulation and Verification via Deep Reinforcement Learning Algorithms." Journal of Physics: Conference Series 2435, no. 1 (February 1, 2023): 012019. http://dx.doi.org/10.1088/1742-6596/2435/1/012019.
Повний текст джерелаDing, Yuhao, Ming Jin, and Javad Lavaei. "Non-stationary Risk-Sensitive Reinforcement Learning: Near-Optimal Dynamic Regret, Adaptive Detection, and Separation Design." Proceedings of the AAAI Conference on Artificial Intelligence 37, no. 6 (June 26, 2023): 7405–13. http://dx.doi.org/10.1609/aaai.v37i6.25901.
Повний текст джерелаSarkar, Soumyadip. "Quantitative Trading using Deep Q Learning." International Journal for Research in Applied Science and Engineering Technology 11, no. 4 (April 30, 2023): 731–38. http://dx.doi.org/10.22214/ijraset.2023.50170.
Повний текст джерелаZhang, Ningyan. "Analysis of reinforce learning in medical treatment." Applied and Computational Engineering 5, no. 1 (June 14, 2023): 48–53. http://dx.doi.org/10.54254/2755-2721/5/20230527.
Повний текст джерелаPuspitasari, Annisa Anggun, and Byung Moo Lee. "A Survey on Reinforcement Learning for Reconfigurable Intelligent Surfaces in Wireless Communications." Sensors 23, no. 5 (February 24, 2023): 2554. http://dx.doi.org/10.3390/s23052554.
Повний текст джерелаDelipetrev, Blagoj, Andreja Jonoski, and Dimitri P. Solomatine. "A novel nested stochastic dynamic programming (nSDP) and nested reinforcement learning (nRL) algorithm for multipurpose reservoir optimization." Journal of Hydroinformatics 19, no. 1 (September 17, 2016): 47–61. http://dx.doi.org/10.2166/hydro.2016.243.
Повний текст джерелаWang, Mengmei. "Optimizing Multitask Assignment of Internet of Things Devices by Reinforcement Learning in Mobile Crowdsensing Scenes." Security and Communication Networks 2022 (August 17, 2022): 1–10. http://dx.doi.org/10.1155/2022/6202237.
Повний текст джерелаГайнетдинов, А. Ф. "NeRF IN REINFORCEMENT LEARNING FOR IMAGE RECOGNITION." Южно-Сибирский научный вестник, no. 2(48) (April 30, 2023): 63–72. http://dx.doi.org/10.25699/sssb.2023.48.2.011.
Повний текст джерелаNicola, Marcel, and Claudiu-Ionel Nicola. "Improvement of Linear and Nonlinear Control for PMSM Using Computational Intelligence and Reinforcement Learning." Mathematics 10, no. 24 (December 9, 2022): 4667. http://dx.doi.org/10.3390/math10244667.
Повний текст джерелаYou, Haoyi, Beichen Yu, Haiming Jin, Zhaoxing Yang, and Jiahui Sun. "User-Oriented Robust Reinforcement Learning." Proceedings of the AAAI Conference on Artificial Intelligence 37, no. 12 (June 26, 2023): 15269–77. http://dx.doi.org/10.1609/aaai.v37i12.26781.
Повний текст джерелаYang, Bin, Muhammad Haseeb Arshad, and Qing Zhao. "Packet-Level and Flow-Level Network Intrusion Detection Based on Reinforcement Learning and Adversarial Training." Algorithms 15, no. 12 (November 30, 2022): 453. http://dx.doi.org/10.3390/a15120453.
Повний текст джерела