Literatura académica sobre el tema "Sparsely rewarded environments"
Crea una cita precisa en los estilos APA, MLA, Chicago, Harvard y otros
Consulte las listas temáticas de artículos, libros, tesis, actas de conferencias y otras fuentes académicas sobre el tema "Sparsely rewarded environments".
Junto a cada fuente en la lista de referencias hay un botón "Agregar a la bibliografía". Pulsa este botón, y generaremos automáticamente la referencia bibliográfica para la obra elegida en el estilo de cita que necesites: APA, MLA, Harvard, Vancouver, Chicago, etc.
También puede descargar el texto completo de la publicación académica en formato pdf y leer en línea su resumen siempre que esté disponible en los metadatos.
Artículos de revistas sobre el tema "Sparsely rewarded environments"
Dubey, Rachit, Thomas L. Griffiths y Peter Dayan. "The pursuit of happiness: A reinforcement learning perspective on habituation and comparisons". PLOS Computational Biology 18, n.º 8 (4 de agosto de 2022): e1010316. http://dx.doi.org/10.1371/journal.pcbi.1010316.
Texto completoShi, Xiaoping, Shiqi Zou, Shenmin Song y Rui Guo. "A multi-objective sparse evolutionary framework for large-scale weapon target assignment based on a reward strategy". Journal of Intelligent & Fuzzy Systems 40, n.º 5 (22 de abril de 2021): 10043–61. http://dx.doi.org/10.3233/jifs-202679.
Texto completoSakamoto, Yuma y Kentarou Kurashige. "Self-Generating Evaluations for Robot’s Autonomy Based on Sensor Input". Machines 11, n.º 9 (6 de septiembre de 2023): 892. http://dx.doi.org/10.3390/machines11090892.
Texto completoParisi, Simone, Davide Tateo, Maximilian Hensel, Carlo D’Eramo, Jan Peters y Joni Pajarinen. "Long-Term Visitation Value for Deep Exploration in Sparse-Reward Reinforcement Learning". Algorithms 15, n.º 3 (28 de febrero de 2022): 81. http://dx.doi.org/10.3390/a15030081.
Texto completoMguni, David, Taher Jafferjee, Jianhong Wang, Nicolas Perez-Nieves, Wenbin Song, Feifei Tong, Matthew Taylor et al. "Learning to Shape Rewards Using a Game of Two Partners". Proceedings of the AAAI Conference on Artificial Intelligence 37, n.º 10 (26 de junio de 2023): 11604–12. http://dx.doi.org/10.1609/aaai.v37i10.26371.
Texto completoForbes, Grant C. y David L. Roberts. "Potential-Based Reward Shaping for Intrinsic Motivation (Student Abstract)". Proceedings of the AAAI Conference on Artificial Intelligence 38, n.º 21 (24 de marzo de 2024): 23488–89. http://dx.doi.org/10.1609/aaai.v38i21.30441.
Texto completoXu, Pei, Junge Zhang, Qiyue Yin, Chao Yu, Yaodong Yang y Kaiqi Huang. "Subspace-Aware Exploration for Sparse-Reward Multi-Agent Tasks". Proceedings of the AAAI Conference on Artificial Intelligence 37, n.º 10 (26 de junio de 2023): 11717–25. http://dx.doi.org/10.1609/aaai.v37i10.26384.
Texto completoKubovčík, Martin, Iveta Dirgová Luptáková y Jiří Pospíchal. "Signal Novelty Detection as an Intrinsic Reward for Robotics". Sensors 23, n.º 8 (14 de abril de 2023): 3985. http://dx.doi.org/10.3390/s23083985.
Texto completoCatacora Ocana, Jim Martin, Roberto Capobianco y Daniele Nardi. "An Overview of Environmental Features that Impact Deep Reinforcement Learning in Sparse-Reward Domains". Journal of Artificial Intelligence Research 76 (26 de abril de 2023): 1181–218. http://dx.doi.org/10.1613/jair.1.14390.
Texto completoZhou, Xiao, Song Zhou, Xingang Mou y Yi He. "Multirobot Collaborative Pursuit Target Robot by Improved MADDPG". Computational Intelligence and Neuroscience 2022 (25 de febrero de 2022): 1–10. http://dx.doi.org/10.1155/2022/4757394.
Texto completoTesis sobre el tema "Sparsely rewarded environments"
Gallouedec, Quentin. "Toward the generalization of reinforcement learning". Electronic Thesis or Diss., Ecully, Ecole centrale de Lyon, 2024. http://www.theses.fr/2024ECDL0013.
Texto completoConventional Reinforcement Learning (RL) involves training a unimodal agent on a single, well-defined task, guided by a gradient-optimized reward signal. This framework does not allow us to envisage a learning agent adapted to real-world problems involving diverse modality streams, multiple tasks, often poorly defined, sometimes not defined at all. Hence, we advocate for transitioning towards a more general framework, aiming to create RL algorithms that more inherently versatile.To advance in this direction, we identify two primary areas of focus. The first aspect involves improving exploration, enabling the agent to learn from the environment with reduced dependence on the reward signal. We present Latent Go-Explore (LGE), an extension of the Go-Explore algorithm. While Go-Explore achieved impressive results, it was constrained by domain-specific knowledge. LGE overcomes these limitations, offering wider applicability within a general framework. In various tested environments, LGE consistently outperforms the baselines, showcasing its enhanced effectiveness and versatility. The second focus is to design a general-purpose agent that can operate in a variety of environments, thus involving a multimodal structure and even transcending the conventional sequential framework of RL. We introduce Jack of All Trades (JAT), a multimodal Transformer-based architecture uniquely tailored to sequential decision tasks. Using a single set of weights, JAT demonstrates robustness and versatility, competing its unique baseline on several RL benchmarks and even showing promising performance on vision and textual tasks. We believe that these two contributions are a valuable step towards a more general approach to RL. In addition, we present other methodological and technical advances that are closely related to our core research question. The first is the introduction of a set of sparsely rewarded simulated robotic environments designed to provide the community with the necessary tools for learning under conditions of low supervision. Notably, three years after its introduction, this contribution has been widely adopted by the community and continues to receive active maintenance and support. On the other hand, we present Open RL Benchmark, our pioneering initiative to provide a comprehensive and fully tracked set of RL experiments, going beyond typical data to include all algorithm-specific and system metrics. This benchmark aims to improve research efficiency by providing out-of-the-box RL data and facilitating accurate reproducibility of experiments. With its community-driven approach, it has quickly become an important resource, documenting over 25,000 runs.These technical and methodological advances, along with the scientific contributions described above, are intended to promote a more general approach to Reinforcement Learning and, we hope, represent a meaningful step toward the eventual development of a more operative RL agent
Hanski, Jari y Kaan Baris Biçak. "An Evaluation of the Unity Machine Learning Agents Toolkit in Dense and Sparse Reward Video Game Environments". Thesis, Uppsala universitet, Institutionen för speldesign, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-444982.
Texto completoCapítulos de libros sobre el tema "Sparsely rewarded environments"
Hensel, Maximilian. "Exploration Methods in Sparse Reward Environments". En Reinforcement Learning Algorithms: Analysis and Applications, 35–45. Cham: Springer International Publishing, 2021. http://dx.doi.org/10.1007/978-3-030-41188-6_4.
Texto completoMoy, Glennn y Slava Shekh. "Evolution Strategies for Sparse Reward Gridworld Environments". En AI 2022: Advances in Artificial Intelligence, 266–78. Cham: Springer International Publishing, 2022. http://dx.doi.org/10.1007/978-3-031-22695-3_19.
Texto completoJeewa, Asad, Anban W. Pillay y Edgar Jembere. "Learning to Generalise in Sparse Reward Navigation Environments". En Artificial Intelligence Research, 85–100. Cham: Springer International Publishing, 2020. http://dx.doi.org/10.1007/978-3-030-66151-9_6.
Texto completoChen, Zhongpeng y Qiang Guan. "Continuous Exploration via Multiple Perspectives in Sparse Reward Environment". En Pattern Recognition and Computer Vision, 57–68. Singapore: Springer Nature Singapore, 2023. http://dx.doi.org/10.1007/978-981-99-8435-0_5.
Texto completoLe, Bang-Giang, Thi-Linh Hoang, Hai-Dang Kieu y Viet-Cuong Ta. "Structural and Compact Latent Representation Learning on Sparse Reward Environments". En Intelligent Information and Database Systems, 40–51. Singapore: Springer Nature Singapore, 2023. http://dx.doi.org/10.1007/978-981-99-5837-5_4.
Texto completoKang, Yongxin, Enmin Zhao, Yifan Zang, Kai Li y Junliang Xing. "Towards a Unified Benchmark for Reinforcement Learning in Sparse Reward Environments". En Communications in Computer and Information Science, 189–201. Singapore: Springer Nature Singapore, 2023. http://dx.doi.org/10.1007/978-981-99-1639-9_16.
Texto completoLiu, Xi, Long Ma, Zhen Chen, Changgang Zheng, Ren Chen, Yong Liao y Shufan Yang. "A Novel State Space Exploration Method for the Sparse-Reward Reinforcement Learning Environment". En Artificial Intelligence XL, 216–21. Cham: Springer Nature Switzerland, 2023. http://dx.doi.org/10.1007/978-3-031-47994-6_18.
Texto completoXie, Zaipeng, Yufeng Zhang, Chentai Qiao y Sitong Shen. "IPERS: Individual Prioritized Experience Replay with Subgoals for Sparse Reward Multi-Agent Reinforcement Learning". En Frontiers in Artificial Intelligence and Applications. IOS Press, 2023. http://dx.doi.org/10.3233/faia230586.
Texto completoShah, Syed Ihtesham Hussain, Antonio Coronato y Muddasar Naeem. "Inverse Reinforcement Learning Based Approach for Investigating Optimal Dynamic Treatment Regime". En Ambient Intelligence and Smart Environments. IOS Press, 2022. http://dx.doi.org/10.3233/aise220052.
Texto completoAbate, Alessandro, Yousif Almulla, James Fox, David Hyland y Michael Wooldridge. "Learning Task Automata for Reinforcement Learning Using Hidden Markov Models". En Frontiers in Artificial Intelligence and Applications. IOS Press, 2023. http://dx.doi.org/10.3233/faia230247.
Texto completoActas de conferencias sobre el tema "Sparsely rewarded environments"
Camacho, Alberto, Rodrigo Toro Icarte, Toryn Q. Klassen, Richard Valenzano y Sheila A. McIlraith. "LTL and Beyond: Formal Languages for Reward Function Specification in Reinforcement Learning". En Twenty-Eighth International Joint Conference on Artificial Intelligence {IJCAI-19}. California: International Joint Conferences on Artificial Intelligence Organization, 2019. http://dx.doi.org/10.24963/ijcai.2019/840.
Texto completoBougie, Nicolas y Ryutaro Ichise. "Towards High-Level Intrinsic Exploration in Reinforcement Learning". En Twenty-Ninth International Joint Conference on Artificial Intelligence and Seventeenth Pacific Rim International Conference on Artificial Intelligence {IJCAI-PRICAI-20}. California: International Joint Conferences on Artificial Intelligence Organization, 2020. http://dx.doi.org/10.24963/ijcai.2020/733.
Texto completoWan, Shanchuan, Yujin Tang, Yingtao Tian y Tomoyuki Kaneko. "DEIR: Efficient and Robust Exploration through Discriminative-Model-Based Episodic Intrinsic Rewards". En Thirty-Second International Joint Conference on Artificial Intelligence {IJCAI-23}. California: International Joint Conferences on Artificial Intelligence Organization, 2023. http://dx.doi.org/10.24963/ijcai.2023/477.
Texto completoNoever, David y Ryerson Burdick. "Puzzle Solving without Search or Human Knowledge: An Unnatural Language Approach". En 9th International Conference on Artificial Intelligence and Applications (AIAPP 2022). Academy and Industry Research Collaboration Center (AIRCC), 2022. http://dx.doi.org/10.5121/csit.2022.120902.
Texto completoChatterjee, Palash, Ashutosh Chapagain, Weizhe Chen y Roni Khardon. "DiSProD: Differentiable Symbolic Propagation of Distributions for Planning". En Thirty-Second International Joint Conference on Artificial Intelligence {IJCAI-23}. California: International Joint Conferences on Artificial Intelligence Organization, 2023. http://dx.doi.org/10.24963/ijcai.2023/591.
Texto completoXu, Pei, Junge Zhang y Kaiqi Huang. "Exploration via Joint Policy Diversity for Sparse-Reward Multi-Agent Tasks". En Thirty-Second International Joint Conference on Artificial Intelligence {IJCAI-23}. California: International Joint Conferences on Artificial Intelligence Organization, 2023. http://dx.doi.org/10.24963/ijcai.2023/37.
Texto completoMemarian, Farzan, Wonjoon Goo, Rudolf Lioutikov, Scott Niekum y Ufuk Topcu. "Self-Supervised Online Reward Shaping in Sparse-Reward Environments". En 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2021. http://dx.doi.org/10.1109/iros51168.2021.9636020.
Texto completoLin, Xingyu, Pengsheng Guo, Carlos Florensa y David Held. "Adaptive Variance for Changing Sparse-Reward Environments". En 2019 International Conference on Robotics and Automation (ICRA). IEEE, 2019. http://dx.doi.org/10.1109/icra.2019.8793650.
Texto completoSeurin, Mathieu, Florian Strub, Philippe Preux y Olivier Pietquin. "Don’t Do What Doesn’t Matter: Intrinsic Motivation with Action Usefulness". En Thirtieth International Joint Conference on Artificial Intelligence {IJCAI-21}. California: International Joint Conferences on Artificial Intelligence Organization, 2021. http://dx.doi.org/10.24963/ijcai.2021/406.
Texto completoJuliani, Arthur, Ahmed Khalifa, Vincent-Pierre Berges, Jonathan Harper, Ervin Teng, Hunter Henry, Adam Crespi, Julian Togelius y Danny Lange. "Obstacle Tower: A Generalization Challenge in Vision, Control, and Planning". En Twenty-Eighth International Joint Conference on Artificial Intelligence {IJCAI-19}. California: International Joint Conferences on Artificial Intelligence Organization, 2019. http://dx.doi.org/10.24963/ijcai.2019/373.
Texto completo