Статті в журналах з теми "Actor-critic methods"
Оформте джерело за APA, MLA, Chicago, Harvard та іншими стилями
Ознайомтеся з топ-50 статей у журналах для дослідження на тему "Actor-critic methods".
Біля кожної праці в переліку літератури доступна кнопка «Додати до бібліографії». Скористайтеся нею – і ми автоматично оформимо бібліографічне посилання на обрану працю в потрібному вам стилі цитування: APA, MLA, «Гарвард», «Чикаго», «Ванкувер» тощо.
Також ви можете завантажити повний текст наукової публікації у форматі «.pdf» та прочитати онлайн анотацію до роботи, якщо відповідні параметри наявні в метаданих.
Переглядайте статті в журналах для різних дисциплін та оформлюйте правильно вашу бібліографію.
Parisi, Simone, Voot Tangkaratt, Jan Peters, and Mohammad Emtiyaz Khan. "TD-regularized actor-critic methods." Machine Learning 108, no. 8-9 (February 21, 2019): 1467–501. http://dx.doi.org/10.1007/s10994-019-05788-0.
Повний текст джерелаWang, Jing, Xuchu Ding, Morteza Lahijanian, Ioannis Ch Paschalidis, and Calin A. Belta. "Temporal logic motion control using actor–critic methods." International Journal of Robotics Research 34, no. 10 (May 26, 2015): 1329–44. http://dx.doi.org/10.1177/0278364915581505.
Повний текст джерелаGrondman, I., M. Vaandrager, L. Busoniu, R. Babuska, and E. Schuitema. "Efficient Model Learning Methods for Actor–Critic Control." IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics) 42, no. 3 (June 2012): 591–602. http://dx.doi.org/10.1109/tsmcb.2011.2170565.
Повний текст джерелаWang, Mingyi, Jianhao Tang, Haoli Zhao, Zhenni Li, and Shengli Xie. "Automatic Compression of Neural Network with Deep Reinforcement Learning Based on Proximal Gradient Method." Mathematics 11, no. 2 (January 9, 2023): 338. http://dx.doi.org/10.3390/math11020338.
Повний текст джерелаSu, Jianyu, Stephen Adams, and Peter Beling. "Value-Decomposition Multi-Agent Actor-Critics." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 13 (May 18, 2021): 11352–60. http://dx.doi.org/10.1609/aaai.v35i13.17353.
Повний текст джерелаSaglam, Baturay, Furkan B. Mutlu, Dogan C. Cicek, and Suleyman S. Kozat. "Actor Prioritized Experience Replay." Journal of Artificial Intelligence Research 78 (November 16, 2023): 639–72. http://dx.doi.org/10.1613/jair.1.14819.
Повний текст джерелаSeo, Kanghyeon, and Jihoon Yang. "Differentially Private Actor and Its Eligibility Trace." Electronics 9, no. 9 (September 10, 2020): 1486. http://dx.doi.org/10.3390/electronics9091486.
Повний текст джерелаSaglam, Baturay, Furkan Mutlu, Dogan Cicek, and Suleyman Kozat. "Actor Prioritized Experience Replay (Abstract Reprint)." Proceedings of the AAAI Conference on Artificial Intelligence 38, no. 20 (March 24, 2024): 22710. http://dx.doi.org/10.1609/aaai.v38i20.30610.
Повний текст джерелаHafez, Muhammad Burhan, Cornelius Weber, Matthias Kerzel, and Stefan Wermter. "Deep intrinsically motivated continuous actor-critic for efficient robotic visuomotor skill learning." Paladyn, Journal of Behavioral Robotics 10, no. 1 (January 1, 2019): 14–29. http://dx.doi.org/10.1515/pjbr-2019-0005.
Повний текст джерелаKong, Minseok, and Jungmin So. "Empirical Analysis of Automated Stock Trading Using Deep Reinforcement Learning." Applied Sciences 13, no. 1 (January 3, 2023): 633. http://dx.doi.org/10.3390/app13010633.
Повний текст джерелаHernandez-Leal, Pablo, Bilal Kartal, and Matthew E. Taylor. "Agent Modeling as Auxiliary Task for Deep Reinforcement Learning." Proceedings of the AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment 15, no. 1 (October 8, 2019): 31–37. http://dx.doi.org/10.1609/aiide.v15i1.5221.
Повний текст джерелаArvindhan, M., and D. Rajesh Kumar. "Adaptive Resource Allocation in Cloud Data Centers using Actor-Critical Deep Reinforcement Learning for Optimized Load Balancing." International Journal on Recent and Innovation Trends in Computing and Communication 11, no. 5s (May 18, 2023): 310–18. http://dx.doi.org/10.17762/ijritcc.v11i5s.6671.
Повний текст джерелаAws, Ahmad, Arkadij Yuschenko, and Vladimir Soloviev. "End-to-end deep reinforcement learning for control of an autonomous underwater robot with an undulating propulsor." Robotics and Technical Cybernetics 12, no. 1 (March 2024): 36–45. http://dx.doi.org/10.31776/rtcj.12105.
Повний текст джерелаZhang, Haifeng, Weizhe Chen, Zeren Huang, Minne Li, Yaodong Yang, Weinan Zhang, and Jun Wang. "Bi-Level Actor-Critic for Multi-Agent Coordination." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 05 (April 3, 2020): 7325–32. http://dx.doi.org/10.1609/aaai.v34i05.6226.
Повний текст джерелаLuo, Ziwei, Jing Hu, Xin Wang, Shu Hu, Bin Kong, Youbing Yin, Qi Song, Xi Wu, and Siwei Lyu. "Stochastic Planner-Actor-Critic for Unsupervised Deformable Image Registration." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 2 (June 28, 2022): 1917–25. http://dx.doi.org/10.1609/aaai.v36i2.20086.
Повний текст джерелаAslani, Mohammad, Mohammad Saadi Mesgari, Stefan Seipel, and Marco Wiering. "Developing adaptive traffic signal control by actor–critic and direct exploration methods." Proceedings of the Institution of Civil Engineers - Transport 172, no. 5 (October 2019): 289–98. http://dx.doi.org/10.1680/jtran.17.00085.
Повний текст джерелаDoya, Kenji. "Reinforcement Learning in Continuous Time and Space." Neural Computation 12, no. 1 (January 1, 2000): 219–45. http://dx.doi.org/10.1162/089976600300015961.
Повний текст джерелаZhu, Qingling, Xiaoqiang Wu, Qiuzhen Lin, and Wei-Neng Chen. "Two-Stage Evolutionary Reinforcement Learning for Enhancing Exploration and Exploitation." Proceedings of the AAAI Conference on Artificial Intelligence 38, no. 18 (March 24, 2024): 20892–900. http://dx.doi.org/10.1609/aaai.v38i18.30079.
Повний текст джерелаJain, Arushi, Gandharv Patil, Ayush Jain, Khimya Khetarpal, and Doina Precup. "Variance Penalized On-Policy and Off-Policy Actor-Critic." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 9 (May 18, 2021): 7899–907. http://dx.doi.org/10.1609/aaai.v35i9.16964.
Повний текст джерелаRyu, Heechang, Hayong Shin, and Jinkyoo Park. "Multi-Agent Actor-Critic with Hierarchical Graph Attention Network." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 05 (April 3, 2020): 7236–43. http://dx.doi.org/10.1609/aaai.v34i05.6214.
Повний текст джерелаShi, Daming, Xudong Guo, Yi Liu, and Wenhui Fan. "Optimal Policy of Multiplayer Poker via Actor-Critic Reinforcement Learning." Entropy 24, no. 6 (May 30, 2022): 774. http://dx.doi.org/10.3390/e24060774.
Повний текст джерелаWang, Hui, Peng Zhang, and Quan Liu. "An Actor-critic Algorithm Using Cross Evaluation of Value Functions." IAES International Journal of Robotics and Automation (IJRA) 7, no. 1 (March 1, 2018): 39. http://dx.doi.org/10.11591/ijra.v7i1.pp39-47.
Повний текст джерелаZhang, Zuozhen, Junzhong Ji, and Jinduo Liu. "MetaRLEC: Meta-Reinforcement Learning for Discovery of Brain Effective Connectivity." Proceedings of the AAAI Conference on Artificial Intelligence 38, no. 9 (March 24, 2024): 10261–69. http://dx.doi.org/10.1609/aaai.v38i9.28892.
Повний текст джерелаZhao, Nan, Zehua Liu, Yiqiang Cheng, and Chao Tian. "Multi-Agent Actor Critic for Channel Allocation in Heterogeneous Networks." International Journal of Mobile Computing and Multimedia Communications 11, no. 1 (January 2020): 23–41. http://dx.doi.org/10.4018/ijmcmc.2020010102.
Повний текст джерелаChen, Haibo, Zhongwei Huang, Xiaorong Zhao, Xiao Liu, Youjun Jiang, Pinyong Geng, Guang Yang, Yewen Cao, and Deqiang Wang. "Policy Optimization of the Power Allocation Algorithm Based on the Actor–Critic Framework in Small Cell Networks." Mathematics 11, no. 7 (April 2, 2023): 1702. http://dx.doi.org/10.3390/math11071702.
Повний текст джерелаYang, Qisong, Thiago D. Simão, Simon H. Tindemans, and Matthijs T. J. Spaan. "WCSAC: Worst-Case Soft Actor Critic for Safety-Constrained Reinforcement Learning." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 12 (May 18, 2021): 10639–46. http://dx.doi.org/10.1609/aaai.v35i12.17272.
Повний текст джерелаWang, Zhihai, Jie Wang, Qi Zhou, Bin Li, and Houqiang Li. "Sample-Efficient Reinforcement Learning via Conservative Model-Based Actor-Critic." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 8 (June 28, 2022): 8612–20. http://dx.doi.org/10.1609/aaai.v36i8.20839.
Повний текст джерелаZhong, Shan, Quan Liu, and QiMing Fu. "Efficient Actor-Critic Algorithm with Hierarchical Model Learning and Planning." Computational Intelligence and Neuroscience 2016 (2016): 1–15. http://dx.doi.org/10.1155/2016/4824072.
Повний текст джерелаWu, Zhenning, Yiming Deng, and Lixing Wang. "A Pinning Actor-Critic Structure-Based Algorithm for Sizing Complex-Shaped Depth Profiles in MFL Inspection with High Degree of Freedom." Complexity 2021 (April 23, 2021): 1–12. http://dx.doi.org/10.1155/2021/9995033.
Повний текст джерелаLiang, Kun, Guoqiang Zhang, Jinhui Guo, and Wentao Li. "An Actor-Critic Hierarchical Reinforcement Learning Model for Course Recommendation." Electronics 12, no. 24 (December 8, 2023): 4939. http://dx.doi.org/10.3390/electronics12244939.
Повний текст джерелаKwon, Ki-Young, Keun-Woo Jung, Dong-Su Yang, and Jooyoung Park. "Autonomous Vehicle Path Tracking Based on Natural Gradient Methods." Journal of Advanced Computational Intelligence and Intelligent Informatics 16, no. 7 (November 20, 2012): 888–93. http://dx.doi.org/10.20965/jaciii.2012.p0888.
Повний текст джерелаLi, Yarong. "Sequence Alignment with Q-Learning Based on the Actor-Critic Model." ACM Transactions on Asian and Low-Resource Language Information Processing 20, no. 5 (July 2, 2021): 1–7. http://dx.doi.org/10.1145/3433540.
Повний текст джерелаJiang, Liang, Ying Nan, Yu Zhang, and Zhihan Li. "Anti-Interception Guidance for Hypersonic Glide Vehicle: A Deep Reinforcement Learning Approach." Aerospace 9, no. 8 (August 4, 2022): 424. http://dx.doi.org/10.3390/aerospace9080424.
Повний текст джерелаLikmeta, Amarildo, Matteo Sacco, Alberto Maria Metelli, and Marcello Restelli. "Wasserstein Actor-Critic: Directed Exploration via Optimism for Continuous-Actions Control." Proceedings of the AAAI Conference on Artificial Intelligence 37, no. 7 (June 26, 2023): 8782–90. http://dx.doi.org/10.1609/aaai.v37i7.26056.
Повний текст джерелаShi, Lei, Tian Li, Lin Wei, Yongcai Tao, Cuixia Li, and Yufei Gao. "FASTune: Towards Fast and Stable Database Tuning System with Reinforcement Learning." Electronics 12, no. 10 (May 10, 2023): 2168. http://dx.doi.org/10.3390/electronics12102168.
Повний текст джерелаYu, Zhiwen, Wenjie Zheng, Kaiwen Zeng, Ruifeng Zhao, Yanxu Zhang, and Mengdi Zeng. "Energy optimization management of microgrid using improved soft actor-critic algorithm." International Journal of Renewable Energy Development 13, no. 2 (February 20, 2024): 329–39. http://dx.doi.org/10.61435/ijred.2024.59988.
Повний текст джерелаIsmail, Ahmed, and Mustafa Baysal. "Dynamic Pricing Based on Demand Response Using Actor–Critic Agent Reinforcement Learning." Energies 16, no. 14 (July 19, 2023): 5469. http://dx.doi.org/10.3390/en16145469.
Повний текст джерелаDrechsler, M. Funk, T. A. Fiorentin, and H. Göllinger. "Actor-Critic Traction Control Based on Reinforcement Learning with Open-Loop Training." Modelling and Simulation in Engineering 2021 (December 7, 2021): 1–10. http://dx.doi.org/10.1155/2021/4641450.
Повний текст джерелаWu, Jiying, Zhong Yang, Haoze Zhuo, Changliang Xu, Chi Zhang, Naifeng He, Luwei Liao, and Zhiyong Wang. "A Supervised Reinforcement Learning Algorithm for Controlling Drone Hovering." Drones 8, no. 3 (February 20, 2024): 69. http://dx.doi.org/10.3390/drones8030069.
Повний текст джерелаQian, Tiancheng, Xue Mei, Pengxiang Xu, Kangqi Ge, and Zhelei Qiu. "Filtration network: A frame sampling strategy via deep reinforcement learning for video captioning." Journal of Intelligent & Fuzzy Systems 40, no. 6 (June 21, 2021): 11085–97. http://dx.doi.org/10.3233/jifs-202249.
Повний текст джерелаWang, Xinshui, Ke Meng, Xu Wang, Zhibin Liu, and Yuefeng Ma. "Dynamic User Resource Allocation for Downlink Multicarrier NOMA with an Actor–Critic Method." Energies 16, no. 7 (March 24, 2023): 2984. http://dx.doi.org/10.3390/en16072984.
Повний текст джерелаMelo, Francisco. "Differential Eligibility Vectors for Advantage Updating and Gradient Methods." Proceedings of the AAAI Conference on Artificial Intelligence 25, no. 1 (August 4, 2011): 441–46. http://dx.doi.org/10.1609/aaai.v25i1.7938.
Повний текст джерелаLyu, Xueguang, Andrea Baisero, Yuchen Xiao, Brett Daley, and Christopher Amato. "On Centralized Critics in Multi-Agent Reinforcement Learning." Journal of Artificial Intelligence Research 77 (May 31, 2023): 295–354. http://dx.doi.org/10.1613/jair.1.14386.
Повний текст джерелаZhao, Mingjun, Haijiang Wu, Di Niu, and Xiaoli Wang. "Reinforced Curriculum Learning on Pre-Trained Neural Machine Translation Models." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 05 (April 3, 2020): 9652–59. http://dx.doi.org/10.1609/aaai.v34i05.6513.
Повний текст джерелаZhao, Jun, Qingliang Zeng, and Bin Guo. "Adaptive Critic Learning-Based Robust Control of Systems with Uncertain Dynamics." Computational Intelligence and Neuroscience 2021 (November 16, 2021): 1–8. http://dx.doi.org/10.1155/2021/2952115.
Повний текст джерелаYue, Longfei, Rennong Yang, Jialiang Zuo, Mengda Yan, Xiaoru Zhao, and Maolong Lv. "Factored Multi-Agent Soft Actor-Critic for Cooperative Multi-Target Tracking of UAV Swarms." Drones 7, no. 3 (February 22, 2023): 150. http://dx.doi.org/10.3390/drones7030150.
Повний текст джерелаZhou, Kun, Wenyong Wang, Teng Hu, and Kai Deng. "Application of Improved Asynchronous Advantage Actor Critic Reinforcement Learning Model on Anomaly Detection." Entropy 23, no. 3 (February 25, 2021): 274. http://dx.doi.org/10.3390/e23030274.
Повний текст джерелаLu, Junqi, Xinning Wu, Su Cao, Xiangke Wang, and Huangchao Yu. "An Implementation of Actor-Critic Algorithm on Spiking Neural Network Using Temporal Coding Method." Applied Sciences 12, no. 20 (October 16, 2022): 10430. http://dx.doi.org/10.3390/app122010430.
Повний текст джерелаOh, Sang Ho, Jeongyoon Kim, Jae Hoon Nah, and Jongyoul Park. "Employing Deep Reinforcement Learning to Cyber-Attack Simulation for Enhancing Cybersecurity." Electronics 13, no. 3 (January 30, 2024): 555. http://dx.doi.org/10.3390/electronics13030555.
Повний текст джерелаSun, Zhiyao, and Guifen Chen. "Enhancing Heterogeneous Network Performance: Advanced Content Popularity Prediction and Efficient Caching." Electronics 13, no. 4 (February 18, 2024): 794. http://dx.doi.org/10.3390/electronics13040794.
Повний текст джерела