Articoli di riviste sul tema "Actor-critic methods"
Cita una fonte nei formati APA, MLA, Chicago, Harvard e in molti altri stili
Vedi i top-50 articoli di riviste per l'attività di ricerca sul tema "Actor-critic methods".
Accanto a ogni fonte nell'elenco di riferimenti c'è un pulsante "Aggiungi alla bibliografia". Premilo e genereremo automaticamente la citazione bibliografica dell'opera scelta nello stile citazionale di cui hai bisogno: APA, MLA, Harvard, Chicago, Vancouver ecc.
Puoi anche scaricare il testo completo della pubblicazione scientifica nel formato .pdf e leggere online l'abstract (il sommario) dell'opera se è presente nei metadati.
Vedi gli articoli di riviste di molte aree scientifiche e compila una bibliografia corretta.
Parisi, Simone, Voot Tangkaratt, Jan Peters e Mohammad Emtiyaz Khan. "TD-regularized actor-critic methods". Machine Learning 108, n. 8-9 (21 febbraio 2019): 1467–501. http://dx.doi.org/10.1007/s10994-019-05788-0.
Wang, Jing, Xuchu Ding, Morteza Lahijanian, Ioannis Ch Paschalidis e Calin A. Belta. "Temporal logic motion control using actor–critic methods". International Journal of Robotics Research 34, n. 10 (26 maggio 2015): 1329–44. http://dx.doi.org/10.1177/0278364915581505.
Grondman, I., M. Vaandrager, L. Busoniu, R. Babuska e E. Schuitema. "Efficient Model Learning Methods for Actor–Critic Control". IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics) 42, n. 3 (giugno 2012): 591–602. http://dx.doi.org/10.1109/tsmcb.2011.2170565.
Wang, Mingyi, Jianhao Tang, Haoli Zhao, Zhenni Li e Shengli Xie. "Automatic Compression of Neural Network with Deep Reinforcement Learning Based on Proximal Gradient Method". Mathematics 11, n. 2 (9 gennaio 2023): 338. http://dx.doi.org/10.3390/math11020338.
Su, Jianyu, Stephen Adams e Peter Beling. "Value-Decomposition Multi-Agent Actor-Critics". Proceedings of the AAAI Conference on Artificial Intelligence 35, n. 13 (18 maggio 2021): 11352–60. http://dx.doi.org/10.1609/aaai.v35i13.17353.
Saglam, Baturay, Furkan B. Mutlu, Dogan C. Cicek e Suleyman S. Kozat. "Actor Prioritized Experience Replay". Journal of Artificial Intelligence Research 78 (16 novembre 2023): 639–72. http://dx.doi.org/10.1613/jair.1.14819.
Seo, Kanghyeon, e Jihoon Yang. "Differentially Private Actor and Its Eligibility Trace". Electronics 9, n. 9 (10 settembre 2020): 1486. http://dx.doi.org/10.3390/electronics9091486.
Saglam, Baturay, Furkan Mutlu, Dogan Cicek e Suleyman Kozat. "Actor Prioritized Experience Replay (Abstract Reprint)". Proceedings of the AAAI Conference on Artificial Intelligence 38, n. 20 (24 marzo 2024): 22710. http://dx.doi.org/10.1609/aaai.v38i20.30610.
Hafez, Muhammad Burhan, Cornelius Weber, Matthias Kerzel e Stefan Wermter. "Deep intrinsically motivated continuous actor-critic for efficient robotic visuomotor skill learning". Paladyn, Journal of Behavioral Robotics 10, n. 1 (1 gennaio 2019): 14–29. http://dx.doi.org/10.1515/pjbr-2019-0005.
Kong, Minseok, e Jungmin So. "Empirical Analysis of Automated Stock Trading Using Deep Reinforcement Learning". Applied Sciences 13, n. 1 (3 gennaio 2023): 633. http://dx.doi.org/10.3390/app13010633.
Hernandez-Leal, Pablo, Bilal Kartal e Matthew E. Taylor. "Agent Modeling as Auxiliary Task for Deep Reinforcement Learning". Proceedings of the AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment 15, n. 1 (8 ottobre 2019): 31–37. http://dx.doi.org/10.1609/aiide.v15i1.5221.
Arvindhan, M., e D. Rajesh Kumar. "Adaptive Resource Allocation in Cloud Data Centers using Actor-Critical Deep Reinforcement Learning for Optimized Load Balancing". International Journal on Recent and Innovation Trends in Computing and Communication 11, n. 5s (18 maggio 2023): 310–18. http://dx.doi.org/10.17762/ijritcc.v11i5s.6671.
Aws, Ahmad, Arkadij Yuschenko e Vladimir Soloviev. "End-to-end deep reinforcement learning for control of an autonomous underwater robot with an undulating propulsor". Robotics and Technical Cybernetics 12, n. 1 (marzo 2024): 36–45. http://dx.doi.org/10.31776/rtcj.12105.
Zhang, Haifeng, Weizhe Chen, Zeren Huang, Minne Li, Yaodong Yang, Weinan Zhang e Jun Wang. "Bi-Level Actor-Critic for Multi-Agent Coordination". Proceedings of the AAAI Conference on Artificial Intelligence 34, n. 05 (3 aprile 2020): 7325–32. http://dx.doi.org/10.1609/aaai.v34i05.6226.
Luo, Ziwei, Jing Hu, Xin Wang, Shu Hu, Bin Kong, Youbing Yin, Qi Song, Xi Wu e Siwei Lyu. "Stochastic Planner-Actor-Critic for Unsupervised Deformable Image Registration". Proceedings of the AAAI Conference on Artificial Intelligence 36, n. 2 (28 giugno 2022): 1917–25. http://dx.doi.org/10.1609/aaai.v36i2.20086.
Aslani, Mohammad, Mohammad Saadi Mesgari, Stefan Seipel e Marco Wiering. "Developing adaptive traffic signal control by actor–critic and direct exploration methods". Proceedings of the Institution of Civil Engineers - Transport 172, n. 5 (ottobre 2019): 289–98. http://dx.doi.org/10.1680/jtran.17.00085.
Doya, Kenji. "Reinforcement Learning in Continuous Time and Space". Neural Computation 12, n. 1 (1 gennaio 2000): 219–45. http://dx.doi.org/10.1162/089976600300015961.
Zhu, Qingling, Xiaoqiang Wu, Qiuzhen Lin e Wei-Neng Chen. "Two-Stage Evolutionary Reinforcement Learning for Enhancing Exploration and Exploitation". Proceedings of the AAAI Conference on Artificial Intelligence 38, n. 18 (24 marzo 2024): 20892–900. http://dx.doi.org/10.1609/aaai.v38i18.30079.
Jain, Arushi, Gandharv Patil, Ayush Jain, Khimya Khetarpal e Doina Precup. "Variance Penalized On-Policy and Off-Policy Actor-Critic". Proceedings of the AAAI Conference on Artificial Intelligence 35, n. 9 (18 maggio 2021): 7899–907. http://dx.doi.org/10.1609/aaai.v35i9.16964.
Ryu, Heechang, Hayong Shin e Jinkyoo Park. "Multi-Agent Actor-Critic with Hierarchical Graph Attention Network". Proceedings of the AAAI Conference on Artificial Intelligence 34, n. 05 (3 aprile 2020): 7236–43. http://dx.doi.org/10.1609/aaai.v34i05.6214.
Shi, Daming, Xudong Guo, Yi Liu e Wenhui Fan. "Optimal Policy of Multiplayer Poker via Actor-Critic Reinforcement Learning". Entropy 24, n. 6 (30 maggio 2022): 774. http://dx.doi.org/10.3390/e24060774.
Wang, Hui, Peng Zhang e Quan Liu. "An Actor-critic Algorithm Using Cross Evaluation of Value Functions". IAES International Journal of Robotics and Automation (IJRA) 7, n. 1 (1 marzo 2018): 39. http://dx.doi.org/10.11591/ijra.v7i1.pp39-47.
Zhang, Zuozhen, Junzhong Ji e Jinduo Liu. "MetaRLEC: Meta-Reinforcement Learning for Discovery of Brain Effective Connectivity". Proceedings of the AAAI Conference on Artificial Intelligence 38, n. 9 (24 marzo 2024): 10261–69. http://dx.doi.org/10.1609/aaai.v38i9.28892.
Zhao, Nan, Zehua Liu, Yiqiang Cheng e Chao Tian. "Multi-Agent Actor Critic for Channel Allocation in Heterogeneous Networks". International Journal of Mobile Computing and Multimedia Communications 11, n. 1 (gennaio 2020): 23–41. http://dx.doi.org/10.4018/ijmcmc.2020010102.
Chen, Haibo, Zhongwei Huang, Xiaorong Zhao, Xiao Liu, Youjun Jiang, Pinyong Geng, Guang Yang, Yewen Cao e Deqiang Wang. "Policy Optimization of the Power Allocation Algorithm Based on the Actor–Critic Framework in Small Cell Networks". Mathematics 11, n. 7 (2 aprile 2023): 1702. http://dx.doi.org/10.3390/math11071702.
Yang, Qisong, Thiago D. Simão, Simon H. Tindemans e Matthijs T. J. Spaan. "WCSAC: Worst-Case Soft Actor Critic for Safety-Constrained Reinforcement Learning". Proceedings of the AAAI Conference on Artificial Intelligence 35, n. 12 (18 maggio 2021): 10639–46. http://dx.doi.org/10.1609/aaai.v35i12.17272.
Wang, Zhihai, Jie Wang, Qi Zhou, Bin Li e Houqiang Li. "Sample-Efficient Reinforcement Learning via Conservative Model-Based Actor-Critic". Proceedings of the AAAI Conference on Artificial Intelligence 36, n. 8 (28 giugno 2022): 8612–20. http://dx.doi.org/10.1609/aaai.v36i8.20839.
Zhong, Shan, Quan Liu e QiMing Fu. "Efficient Actor-Critic Algorithm with Hierarchical Model Learning and Planning". Computational Intelligence and Neuroscience 2016 (2016): 1–15. http://dx.doi.org/10.1155/2016/4824072.
Wu, Zhenning, Yiming Deng e Lixing Wang. "A Pinning Actor-Critic Structure-Based Algorithm for Sizing Complex-Shaped Depth Profiles in MFL Inspection with High Degree of Freedom". Complexity 2021 (23 aprile 2021): 1–12. http://dx.doi.org/10.1155/2021/9995033.
Liang, Kun, Guoqiang Zhang, Jinhui Guo e Wentao Li. "An Actor-Critic Hierarchical Reinforcement Learning Model for Course Recommendation". Electronics 12, n. 24 (8 dicembre 2023): 4939. http://dx.doi.org/10.3390/electronics12244939.
Kwon, Ki-Young, Keun-Woo Jung, Dong-Su Yang e Jooyoung Park. "Autonomous Vehicle Path Tracking Based on Natural Gradient Methods". Journal of Advanced Computational Intelligence and Intelligent Informatics 16, n. 7 (20 novembre 2012): 888–93. http://dx.doi.org/10.20965/jaciii.2012.p0888.
Li, Yarong. "Sequence Alignment with Q-Learning Based on the Actor-Critic Model". ACM Transactions on Asian and Low-Resource Language Information Processing 20, n. 5 (2 luglio 2021): 1–7. http://dx.doi.org/10.1145/3433540.
Jiang, Liang, Ying Nan, Yu Zhang e Zhihan Li. "Anti-Interception Guidance for Hypersonic Glide Vehicle: A Deep Reinforcement Learning Approach". Aerospace 9, n. 8 (4 agosto 2022): 424. http://dx.doi.org/10.3390/aerospace9080424.
Likmeta, Amarildo, Matteo Sacco, Alberto Maria Metelli e Marcello Restelli. "Wasserstein Actor-Critic: Directed Exploration via Optimism for Continuous-Actions Control". Proceedings of the AAAI Conference on Artificial Intelligence 37, n. 7 (26 giugno 2023): 8782–90. http://dx.doi.org/10.1609/aaai.v37i7.26056.
Shi, Lei, Tian Li, Lin Wei, Yongcai Tao, Cuixia Li e Yufei Gao. "FASTune: Towards Fast and Stable Database Tuning System with Reinforcement Learning". Electronics 12, n. 10 (10 maggio 2023): 2168. http://dx.doi.org/10.3390/electronics12102168.
Yu, Zhiwen, Wenjie Zheng, Kaiwen Zeng, Ruifeng Zhao, Yanxu Zhang e Mengdi Zeng. "Energy optimization management of microgrid using improved soft actor-critic algorithm". International Journal of Renewable Energy Development 13, n. 2 (20 febbraio 2024): 329–39. http://dx.doi.org/10.61435/ijred.2024.59988.
Ismail, Ahmed, e Mustafa Baysal. "Dynamic Pricing Based on Demand Response Using Actor–Critic Agent Reinforcement Learning". Energies 16, n. 14 (19 luglio 2023): 5469. http://dx.doi.org/10.3390/en16145469.
Drechsler, M. Funk, T. A. Fiorentin e H. Göllinger. "Actor-Critic Traction Control Based on Reinforcement Learning with Open-Loop Training". Modelling and Simulation in Engineering 2021 (7 dicembre 2021): 1–10. http://dx.doi.org/10.1155/2021/4641450.
Wu, Jiying, Zhong Yang, Haoze Zhuo, Changliang Xu, Chi Zhang, Naifeng He, Luwei Liao e Zhiyong Wang. "A Supervised Reinforcement Learning Algorithm for Controlling Drone Hovering". Drones 8, n. 3 (20 febbraio 2024): 69. http://dx.doi.org/10.3390/drones8030069.
Qian, Tiancheng, Xue Mei, Pengxiang Xu, Kangqi Ge e Zhelei Qiu. "Filtration network: A frame sampling strategy via deep reinforcement learning for video captioning". Journal of Intelligent & Fuzzy Systems 40, n. 6 (21 giugno 2021): 11085–97. http://dx.doi.org/10.3233/jifs-202249.
Wang, Xinshui, Ke Meng, Xu Wang, Zhibin Liu e Yuefeng Ma. "Dynamic User Resource Allocation for Downlink Multicarrier NOMA with an Actor–Critic Method". Energies 16, n. 7 (24 marzo 2023): 2984. http://dx.doi.org/10.3390/en16072984.
Melo, Francisco. "Differential Eligibility Vectors for Advantage Updating and Gradient Methods". Proceedings of the AAAI Conference on Artificial Intelligence 25, n. 1 (4 agosto 2011): 441–46. http://dx.doi.org/10.1609/aaai.v25i1.7938.
Lyu, Xueguang, Andrea Baisero, Yuchen Xiao, Brett Daley e Christopher Amato. "On Centralized Critics in Multi-Agent Reinforcement Learning". Journal of Artificial Intelligence Research 77 (31 maggio 2023): 295–354. http://dx.doi.org/10.1613/jair.1.14386.
Zhao, Mingjun, Haijiang Wu, Di Niu e Xiaoli Wang. "Reinforced Curriculum Learning on Pre-Trained Neural Machine Translation Models". Proceedings of the AAAI Conference on Artificial Intelligence 34, n. 05 (3 aprile 2020): 9652–59. http://dx.doi.org/10.1609/aaai.v34i05.6513.
Zhao, Jun, Qingliang Zeng e Bin Guo. "Adaptive Critic Learning-Based Robust Control of Systems with Uncertain Dynamics". Computational Intelligence and Neuroscience 2021 (16 novembre 2021): 1–8. http://dx.doi.org/10.1155/2021/2952115.
Yue, Longfei, Rennong Yang, Jialiang Zuo, Mengda Yan, Xiaoru Zhao e Maolong Lv. "Factored Multi-Agent Soft Actor-Critic for Cooperative Multi-Target Tracking of UAV Swarms". Drones 7, n. 3 (22 febbraio 2023): 150. http://dx.doi.org/10.3390/drones7030150.
Zhou, Kun, Wenyong Wang, Teng Hu e Kai Deng. "Application of Improved Asynchronous Advantage Actor Critic Reinforcement Learning Model on Anomaly Detection". Entropy 23, n. 3 (25 febbraio 2021): 274. http://dx.doi.org/10.3390/e23030274.
Lu, Junqi, Xinning Wu, Su Cao, Xiangke Wang e Huangchao Yu. "An Implementation of Actor-Critic Algorithm on Spiking Neural Network Using Temporal Coding Method". Applied Sciences 12, n. 20 (16 ottobre 2022): 10430. http://dx.doi.org/10.3390/app122010430.
Oh, Sang Ho, Jeongyoon Kim, Jae Hoon Nah e Jongyoul Park. "Employing Deep Reinforcement Learning to Cyber-Attack Simulation for Enhancing Cybersecurity". Electronics 13, n. 3 (30 gennaio 2024): 555. http://dx.doi.org/10.3390/electronics13030555.
Sun, Zhiyao, e Guifen Chen. "Enhancing Heterogeneous Network Performance: Advanced Content Popularity Prediction and Efficient Caching". Electronics 13, n. 4 (18 febbraio 2024): 794. http://dx.doi.org/10.3390/electronics13040794.