Статті в журналах з теми "Sparse Reward"
Оформте джерело за APA, MLA, Chicago, Harvard та іншими стилями
Ознайомтеся з топ-50 статей у журналах для дослідження на тему "Sparse Reward".
Біля кожної праці в переліку літератури доступна кнопка «Додати до бібліографії». Скористайтеся нею – і ми автоматично оформимо бібліографічне посилання на обрану працю в потрібному вам стилі цитування: APA, MLA, «Гарвард», «Чикаго», «Ванкувер» тощо.
Також ви можете завантажити повний текст наукової публікації у форматі «.pdf» та прочитати онлайн анотацію до роботи, якщо відповідні параметри наявні в метаданих.
Переглядайте статті в журналах для різних дисциплін та оформлюйте правильно вашу бібліографію.
Park, Junseok, Yoonsung Kim, Hee bin Yoo, Min Whoo Lee, Kibeom Kim, Won-Seok Choi, Minsu Lee, and Byoung-Tak Zhang. "Unveiling the Significance of Toddler-Inspired Reward Transition in Goal-Oriented Reinforcement Learning." Proceedings of the AAAI Conference on Artificial Intelligence 38, no. 1 (March 24, 2024): 592–600. http://dx.doi.org/10.1609/aaai.v38i1.27815.
Повний текст джерелаXu, Pei, Junge Zhang, Qiyue Yin, Chao Yu, Yaodong Yang, and Kaiqi Huang. "Subspace-Aware Exploration for Sparse-Reward Multi-Agent Tasks." Proceedings of the AAAI Conference on Artificial Intelligence 37, no. 10 (June 26, 2023): 11717–25. http://dx.doi.org/10.1609/aaai.v37i10.26384.
Повний текст джерелаMguni, David, Taher Jafferjee, Jianhong Wang, Nicolas Perez-Nieves, Wenbin Song, Feifei Tong, Matthew Taylor, et al. "Learning to Shape Rewards Using a Game of Two Partners." Proceedings of the AAAI Conference on Artificial Intelligence 37, no. 10 (June 26, 2023): 11604–12. http://dx.doi.org/10.1609/aaai.v37i10.26371.
Повний текст джерелаMeng, Fanxiao. "Research on Multi-agent Sparse Reward Problem." Highlights in Science, Engineering and Technology 85 (March 13, 2024): 96–103. http://dx.doi.org/10.54097/er0mx710.
Повний текст джерелаZuo, Guoyu, Qishen Zhao, Jiahao Lu, and Jiangeng Li. "Efficient hindsight reinforcement learning using demonstrations for robotic tasks with sparse rewards." International Journal of Advanced Robotic Systems 17, no. 1 (January 1, 2020): 172988141989834. http://dx.doi.org/10.1177/1729881419898342.
Повний текст джерелаVelasquez, Alvaro, Brett Bissey, Lior Barak, Andre Beckus, Ismail Alkhouri, Daniel Melcer, and George Atia. "Dynamic Automaton-Guided Reward Shaping for Monte Carlo Tree Search." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 13 (May 18, 2021): 12015–23. http://dx.doi.org/10.1609/aaai.v35i13.17427.
Повний текст джерелаCorazza, Jan, Ivan Gavran, and Daniel Neider. "Reinforcement Learning with Stochastic Reward Machines." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 6 (June 28, 2022): 6429–36. http://dx.doi.org/10.1609/aaai.v36i6.20594.
Повний текст джерелаGaina, Raluca D., Simon M. Lucas, and Diego Pérez-Liébana. "Tackling Sparse Rewards in Real-Time Games with Statistical Forward Planning Methods." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 1691–98. http://dx.doi.org/10.1609/aaai.v33i01.33011691.
Повний текст джерелаZhou, Xiao, Song Zhou, Xingang Mou, and Yi He. "Multirobot Collaborative Pursuit Target Robot by Improved MADDPG." Computational Intelligence and Neuroscience 2022 (February 25, 2022): 1–10. http://dx.doi.org/10.1155/2022/4757394.
Повний текст джерелаJiang, Jiechuan, and Zongqing Lu. "Generative Exploration and Exploitation." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 04 (April 3, 2020): 4337–44. http://dx.doi.org/10.1609/aaai.v34i04.5858.
Повний текст джерелаYan Kong, Yan Kong, Yefeng Rui Yan Kong, and Chih-Hsien Hsia Yefeng Rui. "A Deep Reinforcement Learning-Based Approach in Porker Game." 電腦學刊 34, no. 2 (April 2023): 041–51. http://dx.doi.org/10.53106/199115992023043402004.
Повний текст джерелаDann, Michael, Fabio Zambetta, and John Thangarajah. "Deriving Subgoals Autonomously to Accelerate Learning in Sparse Reward Domains." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 881–89. http://dx.doi.org/10.1609/aaai.v33i01.3301881.
Повний текст джерелаBougie, Nicolas, and Ryutaro Ichise. "Skill-based curiosity for intrinsically motivated reinforcement learning." Machine Learning 109, no. 3 (October 10, 2019): 493–512. http://dx.doi.org/10.1007/s10994-019-05845-8.
Повний текст джерелаCatacora Ocana, Jim Martin, Roberto Capobianco, and Daniele Nardi. "An Overview of Environmental Features that Impact Deep Reinforcement Learning in Sparse-Reward Domains." Journal of Artificial Intelligence Research 76 (April 26, 2023): 1181–218. http://dx.doi.org/10.1613/jair.1.14390.
Повний текст джерелаZhu, Yiwen, Yuan Zheng, Wenya Wei, and Zhou Fang. "Enhancing Automated Maneuvering Decisions in UCAV Air Combat Games Using Homotopy-Based Reinforcement Learning." Drones 8, no. 12 (December 13, 2024): 756. https://doi.org/10.3390/drones8120756.
Повний текст джерелаGehring, Clement, Masataro Asai, Rohan Chitnis, Tom Silver, Leslie Kaelbling, Shirin Sohrabi, and Michael Katz. "Reinforcement Learning for Classical Planning: Viewing Heuristics as Dense Reward Generators." Proceedings of the International Conference on Automated Planning and Scheduling 32 (June 13, 2022): 588–96. http://dx.doi.org/10.1609/icaps.v32i1.19846.
Повний текст джерелаXu, Zhe, Ivan Gavran, Yousef Ahmad, Rupak Majumdar, Daniel Neider, Ufuk Topcu, and Bo Wu. "Joint Inference of Reward Machines and Policies for Reinforcement Learning." Proceedings of the International Conference on Automated Planning and Scheduling 30 (June 1, 2020): 590–98. http://dx.doi.org/10.1609/icaps.v30i1.6756.
Повний текст джерелаYe, Chenhao, Wei Zhu, Shiluo Guo, and Jinyin Bai. "DQN-Based Shaped Reward Function Mold for UAV Emergency Communication." Applied Sciences 14, no. 22 (November 14, 2024): 10496. http://dx.doi.org/10.3390/app142210496.
Повний текст джерелаDharmavaram, Akshay, Matthew Riemer, and Shalabh Bhatnagar. "Hierarchical Average Reward Policy Gradient Algorithms (Student Abstract)." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 10 (April 3, 2020): 13777–78. http://dx.doi.org/10.1609/aaai.v34i10.7160.
Повний текст джерелаAbu Bakar, Mohamad Hafiz, Abu Ubaidah Shamsudin, Zubair Adil Soomro, Satoshi Tadokoro, and C. J. Salaan. "FUSION SPARSE AND SHAPING REWARD FUNCTION IN SOFT ACTOR-CRITIC DEEP REINFORCEMENT LEARNING FOR MOBILE ROBOT NAVIGATION." Jurnal Teknologi 86, no. 2 (January 15, 2024): 37–49. http://dx.doi.org/10.11113/jurnalteknologi.v86.20147.
Повний текст джерелаSharip, Zati, Mohd Hafiz Zulkifli, Mohd Nur Farhan Abd Wahab, Zubaidi Johar, and Mohd Zaki Mat Amin. "ASSESSING TROPHIC STATE AND WATER QUALITY OF SMALL LAKES AND PONDS IN PERAK." Jurnal Teknologi 86, no. 2 (January 15, 2024): 51–59. http://dx.doi.org/10.11113/jurnalteknologi.v86.20566.
Повний текст джерелаParisi, Simone, Davide Tateo, Maximilian Hensel, Carlo D’Eramo, Jan Peters, and Joni Pajarinen. "Long-Term Visitation Value for Deep Exploration in Sparse-Reward Reinforcement Learning." Algorithms 15, no. 3 (February 28, 2022): 81. http://dx.doi.org/10.3390/a15030081.
Повний текст джерелаForbes, Grant C., and David L. Roberts. "Potential-Based Reward Shaping for Intrinsic Motivation (Student Abstract)." Proceedings of the AAAI Conference on Artificial Intelligence 38, no. 21 (March 24, 2024): 23488–89. http://dx.doi.org/10.1609/aaai.v38i21.30441.
Повний текст джерелаGuo, Yijie, Qiucheng Wu, and Honglak Lee. "Learning Action Translator for Meta Reinforcement Learning on Sparse-Reward Tasks." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 6 (June 28, 2022): 6792–800. http://dx.doi.org/10.1609/aaai.v36i6.20635.
Повний текст джерелаBooth, Serena, W. Bradley Knox, Julie Shah, Scott Niekum, Peter Stone, and Alessandro Allievi. "The Perils of Trial-and-Error Reward Design: Misdesign through Overfitting and Invalid Task Specifications." Proceedings of the AAAI Conference on Artificial Intelligence 37, no. 5 (June 26, 2023): 5920–29. http://dx.doi.org/10.1609/aaai.v37i5.25733.
Повний текст джерелаLinke, Cam, Nadia M. Ady, Martha White, Thomas Degris, and Adam White. "Adapting Behavior via Intrinsic Reward: A Survey and Empirical Study." Journal of Artificial Intelligence Research 69 (December 14, 2020): 1287–332. http://dx.doi.org/10.1613/jair.1.12087.
Повний текст джерелаVelasquez, Alvaro, Brett Bissey, Lior Barak, Daniel Melcer, Andre Beckus, Ismail Alkhouri, and George Atia. "Multi-Agent Tree Search with Dynamic Reward Shaping." Proceedings of the International Conference on Automated Planning and Scheduling 32 (June 13, 2022): 652–61. http://dx.doi.org/10.1609/icaps.v32i1.19854.
Повний текст джерелаSorg, Jonathan, Satinder Singh, and Richard Lewis. "Optimal Rewards versus Leaf-Evaluation Heuristics in Planning Agents." Proceedings of the AAAI Conference on Artificial Intelligence 25, no. 1 (August 4, 2011): 465–70. http://dx.doi.org/10.1609/aaai.v25i1.7931.
Повний текст джерелаYin, Haiyan, Jianda Chen, Sinno Jialin Pan, and Sebastian Tschiatschek. "Sequential Generative Exploration Model for Partially Observable Reinforcement Learning." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 12 (May 18, 2021): 10700–10708. http://dx.doi.org/10.1609/aaai.v35i12.17279.
Повний текст джерелаHasanbeig, Mohammadhosein, Natasha Yogananda Jeppu, Alessandro Abate, Tom Melham, and Daniel Kroening. "DeepSynth: Automata Synthesis for Automatic Task Segmentation in Deep Reinforcement Learning." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 9 (May 18, 2021): 7647–56. http://dx.doi.org/10.1609/aaai.v35i9.16935.
Повний текст джерелаHasanbeig, Hosein, Natasha Yogananda Jeppu, Alessandro Abate, Tom Melham, and Daniel Kroening. "Symbolic Task Inference in Deep Reinforcement Learning." Journal of Artificial Intelligence Research 80 (July 23, 2024): 1099–137. http://dx.doi.org/10.1613/jair.1.14063.
Повний текст джерелаJiang, Nan, Sheng Jin, and Changshui Zhang. "Hierarchical automatic curriculum learning: Converting a sparse reward navigation task into dense reward." Neurocomputing 360 (September 2019): 265–78. http://dx.doi.org/10.1016/j.neucom.2019.06.024.
Повний текст джерелаJin, Tianyuan, Hao-Lun Hsu, William Chang, and Pan Xu. "Finite-Time Frequentist Regret Bounds of Multi-Agent Thompson Sampling on Sparse Hypergraphs." Proceedings of the AAAI Conference on Artificial Intelligence 38, no. 11 (March 24, 2024): 12956–64. http://dx.doi.org/10.1609/aaai.v38i11.29193.
Повний текст джерелаMa, Ang, Yanhua Yu, Chuan Shi, Shuai Zhen, Liang Pang, and Tat-Seng Chua. "PMHR: Path-Based Multi-Hop Reasoning Incorporating Rule-Enhanced Reinforcement Learning and KG Embeddings." Electronics 13, no. 23 (December 9, 2024): 4847. https://doi.org/10.3390/electronics13234847.
Повний текст джерелаWei, Tianqi, Qinghai Guo, and Barbara Webb. "Learning with sparse reward in a gap junction network inspired by the insect mushroom body." PLOS Computational Biology 20, no. 5 (May 23, 2024): e1012086. http://dx.doi.org/10.1371/journal.pcbi.1012086.
Повний текст джерелаKang, Yongxin, Enmin Zhao, Kai Li, and Junliang Xing. "Exploration via State influence Modeling." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 9 (May 18, 2021): 8047–54. http://dx.doi.org/10.1609/aaai.v35i9.16981.
Повний текст джерелаSakamoto, Yuma, and Kentarou Kurashige. "Self-Generating Evaluations for Robot’s Autonomy Based on Sensor Input." Machines 11, no. 9 (September 6, 2023): 892. http://dx.doi.org/10.3390/machines11090892.
Повний текст джерелаMorrison, Sara E., Vincent B. McGinty, Johann du Hoffmann, and Saleem M. Nicola. "Limbic-motor integration by neural excitations and inhibitions in the nucleus accumbens." Journal of Neurophysiology 118, no. 5 (November 1, 2017): 2549–67. http://dx.doi.org/10.1152/jn.00465.2017.
Повний текст джерелаHan, Ziyao, Fan Yi, and Kazuhiro Ohkura. "Collective Transport Behavior in a Robotic Swarm with Hierarchical Imitation Learning." Journal of Robotics and Mechatronics 36, no. 3 (June 20, 2024): 538–45. http://dx.doi.org/10.20965/jrm.2024.p0538.
Повний текст джерелаTang, Wanxing, Chuang Cheng, Haiping Ai, and Li Chen. "Dual-Arm Robot Trajectory Planning Based on Deep Reinforcement Learning under Complex Environment." Micromachines 13, no. 4 (March 31, 2022): 564. http://dx.doi.org/10.3390/mi13040564.
Повний текст джерелаXu, Xibao, Yushen Chen, and Chengchao Bai. "Deep Reinforcement Learning-Based Accurate Control of Planetary Soft Landing." Sensors 21, no. 23 (December 6, 2021): 8161. http://dx.doi.org/10.3390/s21238161.
Повний текст джерелаSong, Qingpeng, Yuansheng Liu, Ming Lu, Jun Zhang, Han Qi, Ziyu Wang, and Zijian Liu. "Autonomous Driving Decision Control Based on Improved Proximal Policy Optimization Algorithm." Applied Sciences 13, no. 11 (May 24, 2023): 6400. http://dx.doi.org/10.3390/app13116400.
Повний текст джерелаPotjans, Wiebke, Abigail Morrison, and Markus Diesmann. "A Spiking Neural Network Model of an Actor-Critic Learning Agent." Neural Computation 21, no. 2 (February 2009): 301–39. http://dx.doi.org/10.1162/neco.2008.08-07-593.
Повний текст джерелаKim, MyeongSeop, and Jung-Su Kim. "Policy-based Deep Reinforcement Learning for Sparse Reward Environment." Transactions of The Korean Institute of Electrical Engineers 70, no. 3 (March 31, 2021): 506–14. http://dx.doi.org/10.5370/kiee.2021.70.3.506.
Повний текст джерелаDai, Tianhong, Hengyan Liu, and Anil Anthony Bharath. "Episodic Self-Imitation Learning with Hindsight." Electronics 9, no. 10 (October 21, 2020): 1742. http://dx.doi.org/10.3390/electronics9101742.
Повний текст джерелаKubovčík, Martin, Iveta Dirgová Luptáková, and Jiří Pospíchal. "Signal Novelty Detection as an Intrinsic Reward for Robotics." Sensors 23, no. 8 (April 14, 2023): 3985. http://dx.doi.org/10.3390/s23083985.
Повний текст джерелаLiu, Yushen. "On the Performance of the Minimax Optimal Strategy in the Stochastic Case of Logistic Bandits." Applied and Computational Engineering 83, no. 1 (October 31, 2024): 130–39. http://dx.doi.org/10.54254/2755-2721/83/2024glg0072.
Повний текст джерелаAlkaff, Muhammad, Abdullah Basuhail, and Yuslena Sari. "Optimizing Water Use in Maize Irrigation with Reinforcement Learning." Mathematics 13, no. 4 (February 11, 2025): 595. https://doi.org/10.3390/math13040595.
Повний текст джерелаde Hauwere, Yann-Michaël, Sam Devlin, Daniel Kudenko, and Ann Nowé. "Context-sensitive reward shaping for sparse interaction multi-agent systems." Knowledge Engineering Review 31, no. 1 (January 2016): 59–76. http://dx.doi.org/10.1017/s0269888915000193.
Повний текст джерелаWang, Xusheng, Jiexin Xie, Shijie Guo, Yue Li, Pengfei Sun, and Zhongxue Gan. "Deep reinforcement learning-based rehabilitation robot trajectory planning with optimized reward functions." Advances in Mechanical Engineering 13, no. 12 (December 2021): 168781402110670. http://dx.doi.org/10.1177/16878140211067011.
Повний текст джерела