Artykuły w czasopismach na temat „Actor-critic algorithm”
Utwórz poprawne odniesienie w stylach APA, MLA, Chicago, Harvard i wielu innych
Sprawdź 50 najlepszych artykułów w czasopismach naukowych na temat „Actor-critic algorithm”.
Przycisk „Dodaj do bibliografii” jest dostępny obok każdej pracy w bibliografii. Użyj go – a my automatycznie utworzymy odniesienie bibliograficzne do wybranej pracy w stylu cytowania, którego potrzebujesz: APA, MLA, Harvard, Chicago, Vancouver itp.
Możesz również pobrać pełny tekst publikacji naukowej w formacie „.pdf” i przeczytać adnotację do pracy online, jeśli odpowiednie parametry są dostępne w metadanych.
Przeglądaj artykuły w czasopismach z różnych dziedzin i twórz odpowiednie bibliografie.
Wang, Jing, i Ioannis Ch Paschalidis. "An Actor-Critic Algorithm With Second-Order Actor and Critic". IEEE Transactions on Automatic Control 62, nr 6 (czerwiec 2017): 2689–703. http://dx.doi.org/10.1109/tac.2016.2616384.
Pełny tekst źródłaZheng, Liyuan, Tanner Fiez, Zane Alumbaugh, Benjamin Chasnov i Lillian J. Ratliff. "Stackelberg Actor-Critic: Game-Theoretic Reinforcement Learning Algorithms". Proceedings of the AAAI Conference on Artificial Intelligence 36, nr 8 (28.06.2022): 9217–24. http://dx.doi.org/10.1609/aaai.v36i8.20908.
Pełny tekst źródłaIwaki, Ryo, i Minoru Asada. "Implicit incremental natural actor critic algorithm". Neural Networks 109 (styczeń 2019): 103–12. http://dx.doi.org/10.1016/j.neunet.2018.10.007.
Pełny tekst źródłaKim, Gi-Soo, Jane P. Kim i Hyun-Joon Yang. "Robust Tests in Online Decision-Making". Proceedings of the AAAI Conference on Artificial Intelligence 36, nr 9 (28.06.2022): 10016–24. http://dx.doi.org/10.1609/aaai.v36i9.21240.
Pełny tekst źródłaSergey, Denisov, i Jee-Hyong Lee. "Actor-Critic Algorithm with Transition Cost Estimation". International Journal of Fuzzy Logic and Intelligent Systems 16, nr 4 (25.12.2016): 270–75. http://dx.doi.org/10.5391/ijfis.2016.16.4.270.
Pełny tekst źródłaAhmed, Ayman Elshabrawy M. "Controller parameter tuning using actor-critic algorithm". IOP Conference Series: Materials Science and Engineering 610 (11.10.2019): 012054. http://dx.doi.org/10.1088/1757-899x/610/1/012054.
Pełny tekst źródłaDing, Siyuan, Shengxiang Li, Guangyi Liu, Ou Li, Ke Ke, Yijie Bai i Weiye Chen. "Decentralized Multiagent Actor-Critic Algorithm Based on Message Diffusion". Journal of Sensors 2021 (8.12.2021): 1–14. http://dx.doi.org/10.1155/2021/8739206.
Pełny tekst źródłaHafez, Muhammad Burhan, Cornelius Weber, Matthias Kerzel i Stefan Wermter. "Deep intrinsically motivated continuous actor-critic for efficient robotic visuomotor skill learning". Paladyn, Journal of Behavioral Robotics 10, nr 1 (1.01.2019): 14–29. http://dx.doi.org/10.1515/pjbr-2019-0005.
Pełny tekst źródłaZhang, Haifei, Jian Xu, Jian Zhang i Quan Liu. "Network Architecture for Optimizing Deep Deterministic Policy Gradient Algorithms". Computational Intelligence and Neuroscience 2022 (18.11.2022): 1–10. http://dx.doi.org/10.1155/2022/1117781.
Pełny tekst źródłaJain, Arushi, Gandharv Patil, Ayush Jain, Khimya Khetarpal i Doina Precup. "Variance Penalized On-Policy and Off-Policy Actor-Critic". Proceedings of the AAAI Conference on Artificial Intelligence 35, nr 9 (18.05.2021): 7899–907. http://dx.doi.org/10.1609/aaai.v35i9.16964.
Pełny tekst źródłaHendzel, Zenon, i Marcin Szuster. "Discrete Action Dependant Heuristic Dynamic Programming in Control of a Wheeled Mobile Robot". Solid State Phenomena 164 (czerwiec 2010): 419–24. http://dx.doi.org/10.4028/www.scientific.net/ssp.164.419.
Pełny tekst źródłaSu, Jianyu, Stephen Adams i Peter Beling. "Value-Decomposition Multi-Agent Actor-Critics". Proceedings of the AAAI Conference on Artificial Intelligence 35, nr 13 (18.05.2021): 11352–60. http://dx.doi.org/10.1609/aaai.v35i13.17353.
Pełny tekst źródłaQiu, Shuang, Zhuoran Yang, Jieping Ye i Zhaoran Wang. "On Finite-Time Convergence of Actor-Critic Algorithm". IEEE Journal on Selected Areas in Information Theory 2, nr 2 (czerwiec 2021): 652–64. http://dx.doi.org/10.1109/jsait.2021.3078754.
Pełny tekst źródłaDing, Feng, Guanfeng Ma, Zhikui Chen, Jing Gao i Peng Li. "Averaged Soft Actor-Critic for Deep Reinforcement Learning". Complexity 2021 (1.04.2021): 1–16. http://dx.doi.org/10.1155/2021/6658724.
Pełny tekst źródłaHatakeyama, Hiroyuki, Shingo Mabu, Kotaro Hirasawa i Jinglu Hu. "Genetic Network Programming with Actor-Critic". Journal of Advanced Computational Intelligence and Intelligent Informatics 11, nr 1 (20.01.2007): 79–86. http://dx.doi.org/10.20965/jaciii.2007.p0079.
Pełny tekst źródłaHyeon, Soo-Jong, Tae-Young Kang i Chang-Kyung Ryoo. "A Path Planning for Unmanned Aerial Vehicles Using SAC (Soft Actor Critic) Algorithm". Journal of Institute of Control, Robotics and Systems 28, nr 2 (28.02.2022): 138–45. http://dx.doi.org/10.5302/j.icros.2022.21.0220.
Pełny tekst źródłaLi, Xinzhou, Guifen Chen, Guowei Wu, Zhiyao Sun i Guangjiao Chen. "Research on Multi-Agent D2D Communication Resource Allocation Algorithm Based on A2C". Electronics 12, nr 2 (10.01.2023): 360. http://dx.doi.org/10.3390/electronics12020360.
Pełny tekst źródłaArvindhan, M., i D. Rajesh Kumar. "Adaptive Resource Allocation in Cloud Data Centers using Actor-Critical Deep Reinforcement Learning for Optimized Load Balancing". International Journal on Recent and Innovation Trends in Computing and Communication 11, nr 5s (18.05.2023): 310–18. http://dx.doi.org/10.17762/ijritcc.v11i5s.6671.
Pełny tekst źródłaSeo, Kanghyeon, i Jihoon Yang. "Differentially Private Actor and Its Eligibility Trace". Electronics 9, nr 9 (10.09.2020): 1486. http://dx.doi.org/10.3390/electronics9091486.
Pełny tekst źródłaLiao, Junrong, Shiyue Liu, Qinghe Wu, Jiabin Chen i Fuhua Wei. "PID Control of Permanent Magnet Synchronous Motor Based on Improved Actor-Critic Framework". Journal of Physics: Conference Series 2213, nr 1 (1.03.2022): 012005. http://dx.doi.org/10.1088/1742-6596/2213/1/012005.
Pełny tekst źródłaZhong, Shan, Quan Liu i QiMing Fu. "Efficient Actor-Critic Algorithm with Hierarchical Model Learning and Planning". Computational Intelligence and Neuroscience 2016 (2016): 1–15. http://dx.doi.org/10.1155/2016/4824072.
Pełny tekst źródłaHIROYASU, Tomoyuki, Akiyuki NAKAMURA, Mitsunori MIKI, Masato YOSHIMI i Hisatake YOKOUCHI. "The Sensuous Lighting Control System using Actor-Critic Algorithm". Journal of Japan Society for Fuzzy Theory and Intelligent Informatics 23, nr 4 (2011): 501–12. http://dx.doi.org/10.3156/jsoft.23.501.
Pełny tekst źródłaBorkar, V. S. "An actor-critic algorithm for constrained Markov decision processes". Systems & Control Letters 54, nr 3 (marzec 2005): 207–13. http://dx.doi.org/10.1016/j.sysconle.2004.08.007.
Pełny tekst źródłaLi, Shuang, Yanghui Yan, Ju Ren, Yuezhi Zhou i Yaoxue Zhang. "A Sample-Efficient Actor-Critic Algorithm for Recommendation Diversification". Chinese Journal of Electronics 29, nr 1 (1.01.2020): 89–96. http://dx.doi.org/10.1049/cje.2019.10.004.
Pełny tekst źródłaShi, Wei, Long Chen i Xia Zhu. "Task Offloading Decision-Making Algorithm for Vehicular Edge Computing: A Deep-Reinforcement-Learning-Based Approach". Sensors 23, nr 17 (1.09.2023): 7595. http://dx.doi.org/10.3390/s23177595.
Pełny tekst źródłaZhou, Chengmin, Bingding Huang i Pasi Fränti. "A review of motion planning algorithms for intelligent robots". Journal of Intelligent Manufacturing 33, nr 2 (25.11.2021): 387–424. http://dx.doi.org/10.1007/s10845-021-01867-z.
Pełny tekst źródłaYue, Han, Jiapeng Liu, Dongmei Tian i Qin Zhang. "A Novel Anti-Risk Method for Portfolio Trading Using Deep Reinforcement Learning". Electronics 11, nr 9 (7.05.2022): 1506. http://dx.doi.org/10.3390/electronics11091506.
Pełny tekst źródłaNakamura, Yutaka, Takeshi Mori, Yoichi Tokita, Tomohiro Shibata i Shin Ishii. "Off-Policy Natural Policy Gradient Method for a Biped Walking Using a CPG Controller". Journal of Robotics and Mechatronics 17, nr 6 (20.12.2005): 636–44. http://dx.doi.org/10.20965/jrm.2005.p0636.
Pełny tekst źródłaHwang, Ha Jun, Jaeyeon Jang, Jongkwan Choi, Jung Ho Bae, Sung Ho Kim i Chang Ouk Kim. "Stepwise Soft Actor–Critic for UAV Autonomous Flight Control". Drones 7, nr 9 (24.08.2023): 549. http://dx.doi.org/10.3390/drones7090549.
Pełny tekst źródłaTakano, Toshiaki, Haruhiko Takase, Hiroharu Kawanaka i Shinji Tsuruoka. "Merging with Extraction Method for Transfer Learning in Actor-Critic". Journal of Advanced Computational Intelligence and Intelligent Informatics 15, nr 7 (20.09.2011): 814–21. http://dx.doi.org/10.20965/jaciii.2011.p0814.
Pełny tekst źródłaZhang, Haifeng, Weizhe Chen, Zeren Huang, Minne Li, Yaodong Yang, Weinan Zhang i Jun Wang. "Bi-Level Actor-Critic for Multi-Agent Coordination". Proceedings of the AAAI Conference on Artificial Intelligence 34, nr 05 (3.04.2020): 7325–32. http://dx.doi.org/10.1609/aaai.v34i05.6226.
Pełny tekst źródłaWang, Hui, Peng Zhang i Quan Liu. "An Actor-critic Algorithm Using Cross Evaluation of Value Functions". IAES International Journal of Robotics and Automation (IJRA) 7, nr 1 (1.03.2018): 39. http://dx.doi.org/10.11591/ijra.v7i1.pp39-47.
Pełny tekst źródłaLan, Xuejing, Zhifeng Tan, Tao Zou i Wenbiao Xu. "CACLA-Based Trajectory Tracking Guidance for RLV in Terminal Area Energy Management Phase". Sensors 21, nr 15 (26.07.2021): 5062. http://dx.doi.org/10.3390/s21155062.
Pełny tekst źródłaZhang, Shangtong, i Hengshuai Yao. "ACE: An Actor Ensemble Algorithm for Continuous Control with Tree Search". Proceedings of the AAAI Conference on Artificial Intelligence 33 (17.07.2019): 5789–96. http://dx.doi.org/10.1609/aaai.v33i01.33015789.
Pełny tekst źródłaKIMURA, Hajime, i Shigenobu KOBAYASHI. "An Actor-Critic Algorithm Using a Binary Tree Action Selector". Transactions of the Society of Instrument and Control Engineers 37, nr 12 (2001): 1147–55. http://dx.doi.org/10.9746/sicetr1965.37.1147.
Pełny tekst źródłaZhong, Shan, Jack Tan, Husheng Dong, Xuemei Chen, Shengrong Gong i Zhenjiang Qian. "Modeling-Learning-Based Actor-Critic Algorithm with Gaussian Process Approximator". Journal of Grid Computing 18, nr 2 (18.04.2020): 181–95. http://dx.doi.org/10.1007/s10723-020-09512-4.
Pełny tekst źródłaBorkar, Vivek S., i Vijaymohan R. Konda. "The actor-critic algorithm as multi-time-scale stochastic approximation". Sadhana 22, nr 4 (sierpień 1997): 525–43. http://dx.doi.org/10.1007/bf02745577.
Pełny tekst źródłaItoh, Hideaki, i Kazuyuki Aihara. "Combination of an actor/critic algorithm with goal-directed reasoning". Artificial Life and Robotics 5, nr 4 (grudzień 2001): 233–41. http://dx.doi.org/10.1007/bf02481507.
Pełny tekst źródłaSu, Jie-Ying, Jia-Lin Kang i Shi-Shang Jang. "An Actor-Critic Algorithm for the Stochastic Cutting Stock Problem". Processes 11, nr 4 (13.04.2023): 1203. http://dx.doi.org/10.3390/pr11041203.
Pełny tekst źródłaWu, Zhenning, Yiming Deng i Lixing Wang. "A Pinning Actor-Critic Structure-Based Algorithm for Sizing Complex-Shaped Depth Profiles in MFL Inspection with High Degree of Freedom". Complexity 2021 (23.04.2021): 1–12. http://dx.doi.org/10.1155/2021/9995033.
Pełny tekst źródłaTAHAMI, EHSAN, AMIR HOMAYOUN JAFARI i ALI FALLAH. "APPLICATION OF AN EVOLUTIONARY ACTOR–CRITIC REINFORCEMENT LEARNING METHOD FOR THE CONTROL OF A THREE-LINK MUSCULOSKELETAL ARM DURING A REACHING MOVEMENT". Journal of Mechanics in Medicine and Biology 13, nr 02 (kwiecień 2013): 1350040. http://dx.doi.org/10.1142/s0219519413500401.
Pełny tekst źródłaDoya, Kenji. "Reinforcement Learning in Continuous Time and Space". Neural Computation 12, nr 1 (1.01.2000): 219–45. http://dx.doi.org/10.1162/089976600300015961.
Pełny tekst źródłaChen, Haibo, Zhongwei Huang, Xiaorong Zhao, Xiao Liu, Youjun Jiang, Pinyong Geng, Guang Yang, Yewen Cao i Deqiang Wang. "Policy Optimization of the Power Allocation Algorithm Based on the Actor–Critic Framework in Small Cell Networks". Mathematics 11, nr 7 (2.04.2023): 1702. http://dx.doi.org/10.3390/math11071702.
Pełny tekst źródłaPradhan, Arabinda, Sukant Kishoro Bisoy i Mangal Sain. "Action-Based Load Balancing Technique in Cloud Network Using Actor-Critic-Swarm Optimization". Wireless Communications and Mobile Computing 2022 (30.06.2022): 1–17. http://dx.doi.org/10.1155/2022/6456242.
Pełny tekst źródłaYang, Qisong, Thiago D. Simão, Simon H. Tindemans i Matthijs T. J. Spaan. "WCSAC: Worst-Case Soft Actor Critic for Safety-Constrained Reinforcement Learning". Proceedings of the AAAI Conference on Artificial Intelligence 35, nr 12 (18.05.2021): 10639–46. http://dx.doi.org/10.1609/aaai.v35i12.17272.
Pełny tekst źródłaAvkhimenia, Vadim, Matheus Gemignani, Tim Weis i Petr Musilek. "Deep Reinforcement Learning-Based Operation of Transmission Battery Storage with Dynamic Thermal Line Rating". Energies 15, nr 23 (29.11.2022): 9032. http://dx.doi.org/10.3390/en15239032.
Pełny tekst źródłaSola, Yoann, Gilles Le Chenadec i Benoit Clement. "Simultaneous Control and Guidance of an AUV Based on Soft Actor–Critic". Sensors 22, nr 16 (14.08.2022): 6072. http://dx.doi.org/10.3390/s22166072.
Pełny tekst źródłaWang, Xiao, Zhe Ma, Lei Mao, Kewu Sun, Xuhui Huang, Changchao Fan i Jiake Li. "Accelerating Fuzzy Actor–Critic Learning via Suboptimal Knowledge for a Multi-Agent Tracking Problem". Electronics 12, nr 8 (13.04.2023): 1852. http://dx.doi.org/10.3390/electronics12081852.
Pełny tekst źródłaAli, Hamid, Hammad Majeed, Imran Usman i Khaled A. Almejalli. "Reducing Entropy Overestimation in Soft Actor Critic Using Dual Policy Network". Wireless Communications and Mobile Computing 2021 (10.06.2021): 1–13. http://dx.doi.org/10.1155/2021/9920591.
Pełny tekst źródłaXi, Bao, Rui Wang, Ying-Hao Cai, Tao Lu i Shuo Wang. "A Novel Heterogeneous Actor-critic Algorithm with Recent Emphasizing Replay Memory". International Journal of Automation and Computing 18, nr 4 (23.04.2021): 619–31. http://dx.doi.org/10.1007/s11633-021-1296-x.
Pełny tekst źródła