Artykuły w czasopismach na temat „Safe RL”
Utwórz poprawne odniesienie w stylach APA, MLA, Chicago, Harvard i wielu innych
Sprawdź 50 najlepszych artykułów w czasopismach naukowych na temat „Safe RL”.
Przycisk „Dodaj do bibliografii” jest dostępny obok każdej pracy w bibliografii. Użyj go – a my automatycznie utworzymy odniesienie bibliograficzne do wybranej pracy w stylu cytowania, którego potrzebujesz: APA, MLA, Harvard, Chicago, Vancouver itp.
Możesz również pobrać pełny tekst publikacji naukowej w formacie „.pdf” i przeczytać adnotację do pracy online, jeśli odpowiednie parametry są dostępne w metadanych.
Przeglądaj artykuły w czasopismach z różnych dziedzin i twórz odpowiednie bibliografie.
Carr, Steven, Nils Jansen, Sebastian Junges i Ufuk Topcu. "Safe Reinforcement Learning via Shielding under Partial Observability". Proceedings of the AAAI Conference on Artificial Intelligence 37, nr 12 (26.06.2023): 14748–56. http://dx.doi.org/10.1609/aaai.v37i12.26723.
Pełny tekst źródłaMa, Yecheng Jason, Andrew Shen, Osbert Bastani i Jayaraman Dinesh. "Conservative and Adaptive Penalty for Model-Based Safe Reinforcement Learning". Proceedings of the AAAI Conference on Artificial Intelligence 36, nr 5 (28.06.2022): 5404–12. http://dx.doi.org/10.1609/aaai.v36i5.20478.
Pełny tekst źródłaXu, Haoran, Xianyuan Zhan i Xiangyu Zhu. "Constraints Penalized Q-learning for Safe Offline Reinforcement Learning". Proceedings of the AAAI Conference on Artificial Intelligence 36, nr 8 (28.06.2022): 8753–60. http://dx.doi.org/10.1609/aaai.v36i8.20855.
Pełny tekst źródłaThananjeyan, Brijen, Ashwin Balakrishna, Suraj Nair, Michael Luo, Krishnan Srinivasan, Minho Hwang, Joseph E. Gonzalez, Julian Ibarz, Chelsea Finn i Ken Goldberg. "Recovery RL: Safe Reinforcement Learning With Learned Recovery Zones". IEEE Robotics and Automation Letters 6, nr 3 (lipiec 2021): 4915–22. http://dx.doi.org/10.1109/lra.2021.3070252.
Pełny tekst źródłaSerrano-Cuevas, Jonathan, Eduardo F. Morales i Pablo Hernández-Leal. "Safe reinforcement learning using risk mapping by similarity". Adaptive Behavior 28, nr 4 (18.07.2019): 213–24. http://dx.doi.org/10.1177/1059712319859650.
Pełny tekst źródłaCheng, Richard, Gábor Orosz, Richard M. Murray i Joel W. Burdick. "End-to-End Safe Reinforcement Learning through Barrier Functions for Safety-Critical Continuous Control Tasks". Proceedings of the AAAI Conference on Artificial Intelligence 33 (17.07.2019): 3387–95. http://dx.doi.org/10.1609/aaai.v33i01.33013387.
Pełny tekst źródłaJurj, Sorin Liviu, Dominik Grundt, Tino Werner, Philipp Borchers, Karina Rothemann i Eike Möhlmann. "Increasing the Safety of Adaptive Cruise Control Using Physics-Guided Reinforcement Learning". Energies 14, nr 22 (12.11.2021): 7572. http://dx.doi.org/10.3390/en14227572.
Pełny tekst źródłaSakrihei, Helen. "Using automatic storage for ILL – experiences from the National Repository Library in Norway". Interlending & Document Supply 44, nr 1 (15.02.2016): 14–16. http://dx.doi.org/10.1108/ilds-11-2015-0035.
Pełny tekst źródłaDing, Yuhao, i Javad Lavaei. "Provably Efficient Primal-Dual Reinforcement Learning for CMDPs with Non-stationary Objectives and Constraints". Proceedings of the AAAI Conference on Artificial Intelligence 37, nr 6 (26.06.2023): 7396–404. http://dx.doi.org/10.1609/aaai.v37i6.25900.
Pełny tekst źródłaTubeuf, Carlotta, Felix Birkelbach, Anton Maly i René Hofmann. "Increasing the Flexibility of Hydropower with Reinforcement Learning on a Digital Twin Platform". Energies 16, nr 4 (11.02.2023): 1796. http://dx.doi.org/10.3390/en16041796.
Pełny tekst źródłaYOON, JAE UNG, i JUHONG LEE. "Uncertainty Sequence Modeling Approach for Safe and Effective Autonomous Driving". Korean Institute of Smart Media 11, nr 9 (31.10.2022): 9–20. http://dx.doi.org/10.30693/smj.2022.11.9.9.
Pełny tekst źródłaLin, Xingbin, Deyu Yuan i Xifei Li. "Reinforcement Learning with Dual Safety Policies for Energy Savings in Building Energy Systems". Buildings 13, nr 3 (21.02.2023): 580. http://dx.doi.org/10.3390/buildings13030580.
Pełny tekst źródłaMarchesini, Enrico, Davide Corsi i Alessandro Farinelli. "Exploring Safer Behaviors for Deep Reinforcement Learning". Proceedings of the AAAI Conference on Artificial Intelligence 36, nr 7 (28.06.2022): 7701–9. http://dx.doi.org/10.1609/aaai.v36i7.20737.
Pełny tekst źródłaEgleston, David, Patricia Ann Castelli i Thomas George Marx. "Developing, validating, and testing a model of reflective leadership". Leadership & Organization Development Journal 38, nr 7 (4.09.2017): 886–96. http://dx.doi.org/10.1108/lodj-09-2016-0230.
Pełny tekst źródłaHuh, Gene, i Wonjae Cha. "Development and Clinical Application of Real-Time Light-Guided Vocal Fold Injection". Journal of The Korean Society of Laryngology, Phoniatrics and Logopedics 33, nr 1 (30.04.2022): 1–6. http://dx.doi.org/10.22469/jkslp.2022.33.1.1.
Pełny tekst źródłaRamakrishnan, Ramya, Ece Kamar, Debadeepta Dey, Eric Horvitz i Julie Shah. "Blind Spot Detection for Safe Sim-to-Real Transfer". Journal of Artificial Intelligence Research 67 (4.02.2020): 191–234. http://dx.doi.org/10.1613/jair.1.11436.
Pełny tekst źródłaHao, Hao, Yichen Sun, Xueyun Mei i Yanjun Zhou. "Reverse Logistics Network Design of Electric Vehicle Batteries considering Recall Risk". Mathematical Problems in Engineering 2021 (18.08.2021): 1–16. http://dx.doi.org/10.1155/2021/5518049.
Pełny tekst źródłaRay, Kaustabha, i Ansuman Banerjee. "Horizontal Auto-Scaling for Multi-Access Edge Computing Using Safe Reinforcement Learning". ACM Transactions on Embedded Computing Systems 20, nr 6 (30.11.2021): 1–33. http://dx.doi.org/10.1145/3475991.
Pełny tekst źródłaDelgado, Tomás, Marco Sánchez Sorondo, Víctor Braberman i Sebastián Uchitel. "Exploration Policies for On-the-Fly Controller Synthesis: A Reinforcement Learning Approach". Proceedings of the International Conference on Automated Planning and Scheduling 33, nr 1 (1.07.2023): 569–77. http://dx.doi.org/10.1609/icaps.v33i1.27238.
Pełny tekst źródłaBolster, Lauren, Mark Bosch, Brian Brownbridge i Anurag Saxena. "RAP Trial: Ringer's Lactate and Packed Red Blood Cell Transfusion, An in Vitro Study and Chart Review." Blood 114, nr 22 (20.11.2009): 2105. http://dx.doi.org/10.1182/blood.v114.22.2105.2105.
Pełny tekst źródłaRomey, Aurore, Hussaini G. Ularamu, Abdulnaci Bulut, Syed M. Jamal, Salman Khan, Muhammad Ishaq, Michael Eschbaumer i in. "Field Evaluation of a Safe, Easy, and Low-Cost Protocol for Shipment of Samples from Suspected Cases of Foot-and-Mouth Disease to Diagnostic Laboratories". Transboundary and Emerging Diseases 2023 (5.08.2023): 1–15. http://dx.doi.org/10.1155/2023/9555213.
Pełny tekst źródłaDai, Juntao, Jiaming Ji, Long Yang, Qian Zheng i Gang Pan. "Augmented Proximal Policy Optimization for Safe Reinforcement Learning". Proceedings of the AAAI Conference on Artificial Intelligence 37, nr 6 (26.06.2023): 7288–95. http://dx.doi.org/10.1609/aaai.v37i6.25888.
Pełny tekst źródłaKrstić, Mladen, Giulio Paolo Agnusdei, Pier Paolo Miglietta, Snežana Tadić i Violeta Roso. "Applicability of Industry 4.0 Technologies in the Reverse Logistics: A Circular Economy Approach Based on COmprehensive Distance Based RAnking (COBRA) Method". Sustainability 14, nr 9 (7.05.2022): 5632. http://dx.doi.org/10.3390/su14095632.
Pełny tekst źródłaPrasetyo, Risky Vitria, Abdul Latief Azis i Soegeng Soegijanto. "Comparison of the efficacy and safety of hydroxyethyl starch 130/0.4 and Ringer's lactate in children with grade III dengue hemorrhagic fever". Paediatrica Indonesiana 49, nr 2 (30.04.2009): 97. http://dx.doi.org/10.14238/pi49.2.2009.97-103.
Pełny tekst źródłaBöck, Markus, Julien Malle, Daniel Pasterk, Hrvoje Kukina, Ramin Hasani i Clemens Heitzinger. "Superhuman performance on sepsis MIMIC-III data by distributional reinforcement learning". PLOS ONE 17, nr 11 (3.11.2022): e0275358. http://dx.doi.org/10.1371/journal.pone.0275358.
Pełny tekst źródłaLi, Yue, Xiao Yong Bai, Shi Jie Wang, Luo Yi Qin, Yi Chao Tian i Guang Jie Luo. "Evaluating of the spatial heterogeneity of soil loss tolerance and its effects on erosion risk in the carbonate areas of southern China". Solid Earth 8, nr 3 (29.05.2017): 661–69. http://dx.doi.org/10.5194/se-8-661-2017.
Pełny tekst źródłaKondrup, Flemming, Thomas Jiralerspong, Elaine Lau, Nathan De Lara, Jacob Shkrob, My Duc Tran, Doina Precup i Sumana Basu. "Towards Safe Mechanical Ventilation Treatment Using Deep Offline Reinforcement Learning". Proceedings of the AAAI Conference on Artificial Intelligence 37, nr 13 (26.06.2023): 15696–702. http://dx.doi.org/10.1609/aaai.v37i13.26862.
Pełny tekst źródłaMiyajima, Hirofumi, Noritaka Shigei, Syunki Makino, Hiromi Miyajima, Yohtaro Miyanishi, Shinji Kitagami i Norio Shiratori. "A proposal of privacy preserving reinforcement learning for secure multiparty computation". Artificial Intelligence Research 6, nr 2 (23.05.2017): 57. http://dx.doi.org/10.5430/air.v6n2p57.
Pełny tekst źródłaThananjeyan, Brijen, Ashwin Balakrishna, Ugo Rosolia, Felix Li, Rowan McAllister, Joseph E. Gonzalez, Sergey Levine, Francesco Borrelli i Ken Goldberg. "Safety Augmented Value Estimation From Demonstrations (SAVED): Safe Deep Model-Based RL for Sparse Cost Robotic Tasks". IEEE Robotics and Automation Letters 5, nr 2 (kwiecień 2020): 3612–19. http://dx.doi.org/10.1109/lra.2020.2976272.
Pełny tekst źródłaRen, Tianzhu, Yuanchang Xie i Liming Jiang. "Cooperative Highway Work Zone Merge Control Based on Reinforcement Learning in a Connected and Automated Environment". Transportation Research Record: Journal of the Transportation Research Board 2674, nr 10 (17.07.2020): 363–74. http://dx.doi.org/10.1177/0361198120935873.
Pełny tekst źródłaReda, Ahmad, i József Vásárhelyi. "Design and Implementation of Reinforcement Learning for Automated Driving Compared to Classical MPC Control". Designs 7, nr 1 (29.01.2023): 18. http://dx.doi.org/10.3390/designs7010018.
Pełny tekst źródłaGardille, Arnaud, i Ola Ahmad. "Towards Safe Reinforcement Learning via OOD Dynamics Detection in Autonomous Driving System (Student Abstract)". Proceedings of the AAAI Conference on Artificial Intelligence 37, nr 13 (26.06.2023): 16216–17. http://dx.doi.org/10.1609/aaai.v37i13.26968.
Pełny tekst źródłaFree, David. "In the News". College & Research Libraries News 80, nr 10 (5.11.2019): 541. http://dx.doi.org/10.5860/crln.80.10.541.
Pełny tekst źródłaXu, Xibao, Yushen Chen i Chengchao Bai. "Deep Reinforcement Learning-Based Accurate Control of Planetary Soft Landing". Sensors 21, nr 23 (6.12.2021): 8161. http://dx.doi.org/10.3390/s21238161.
Pełny tekst źródłaSimão, Thiago D., Marnix Suilen i Nils Jansen. "Safe Policy Improvement for POMDPs via Finite-State Controllers". Proceedings of the AAAI Conference on Artificial Intelligence 37, nr 12 (26.06.2023): 15109–17. http://dx.doi.org/10.1609/aaai.v37i12.26763.
Pełny tekst źródłaZhang, Linrui, Qin Zhang, Li Shen, Bo Yuan, Xueqian Wang i Dacheng Tao. "Evaluating Model-Free Reinforcement Learning toward Safety-Critical Tasks". Proceedings of the AAAI Conference on Artificial Intelligence 37, nr 12 (26.06.2023): 15313–21. http://dx.doi.org/10.1609/aaai.v37i12.26786.
Pełny tekst źródłaAngele, Martin K., Nadia Smail, Markus W. Knöferl, Alfred Ayala, William G. Cioffi i Irshad H. Chaudry. "l-Arginine restores splenocyte functions after trauma and hemorrhage potentially by improving splenic blood flow". American Journal of Physiology-Cell Physiology 276, nr 1 (1.01.1999): C145—C151. http://dx.doi.org/10.1152/ajpcell.1999.276.1.c145.
Pełny tekst źródłaStaessens, Tom, Tom Lefebvre i Guillaume Crevecoeur. "Optimizing Cascaded Control of Mechatronic Systems through Constrained Residual Reinforcement Learning". Machines 11, nr 3 (20.03.2023): 402. http://dx.doi.org/10.3390/machines11030402.
Pełny tekst źródłaLv, Kexuan, Xiaofei Pei, Ci Chen i Jie Xu. "A Safe and Efficient Lane Change Decision-Making Strategy of Autonomous Driving Based on Deep Reinforcement Learning". Mathematics 10, nr 9 (5.05.2022): 1551. http://dx.doi.org/10.3390/math10091551.
Pełny tekst źródłaJurj, Sorin Liviu, Tino Werner, Dominik Grundt, Willem Hagemann i Eike Möhlmann. "Towards Safe and Sustainable Autonomous Vehicles Using Environmentally-Friendly Criticality Metrics". Sustainability 14, nr 12 (7.06.2022): 6988. http://dx.doi.org/10.3390/su14126988.
Pełny tekst źródłaMaw, Aye Aye, Maxim Tyan, Tuan Anh Nguyen i Jae-Woo Lee. "iADA*-RL: Anytime Graph-Based Path Planning with Deep Reinforcement Learning for an Autonomous UAV". Applied Sciences 11, nr 9 (27.04.2021): 3948. http://dx.doi.org/10.3390/app11093948.
Pełny tekst źródłaCivetta, Joseph M., i Charles L. Fox. "Advantages of Resuscitation with Balanced Hypertonic Sodium Solution in Disasters". Prehospital and Disaster Medicine 1, S1 (1985): 179–80. http://dx.doi.org/10.1017/s1049023x0004437x.
Pełny tekst źródłaWysocka, B. A., Z. Kassam, G. Lockwood, J. Brierley, L. Dawson i J. Ringash. "Assessment of intra and interfractional organ motion during adjuvant radiochemotherapy in gastric cancer". Journal of Clinical Oncology 25, nr 18_suppl (20.06.2007): 15132. http://dx.doi.org/10.1200/jco.2007.25.18_suppl.15132.
Pełny tekst źródłaNiu, Tong, i Mohit Bansal. "AvgOut: A Simple Output-Probability Measure to Eliminate Dull Responses". Proceedings of the AAAI Conference on Artificial Intelligence 34, nr 05 (3.04.2020): 8560–67. http://dx.doi.org/10.1609/aaai.v34i05.6378.
Pełny tekst źródłaVivek, Kumar, Shah Amiti, Saha Shivshankar i Choudhary Lalit. "Electrolyte and Haemogram changes post large volume liposuction comparing two different tumescent solutions". Indian Journal of Plastic Surgery 47, nr 03 (wrzesień 2014): 386–93. http://dx.doi.org/10.4103/0970-0358.146604.
Pełny tekst źródłaChebbi, Alif, Massimiliano Tazzari, Cristiana Rizzi, Franco Hernan Gomez Tovar, Sara Villa, Silvia Sbaffoni, Mentore Vaccari i Andrea Franzetti. "Burkholderia thailandensis E264 as a promising safe rhamnolipids’ producer towards a sustainable valorization of grape marcs and olive mill pomace". Applied Microbiology and Biotechnology 105, nr 9 (20.04.2021): 3825–42. http://dx.doi.org/10.1007/s00253-021-11292-0.
Pełny tekst źródłaBrown, Jennifer R., Matthew S. Davids, Jordi Rodon, Pau Abrisqueta, Coumaran Egile, Rodrigo Ruiz-Soto i Farrukh Awan. "Update On The Safety and Efficacy Of The Pan Class I PI3K Inhibitor SAR245408 (XL147) In Chronic Lymphocytic Leukemia and Non-Hodgkin’s Lymphoma Patients". Blood 122, nr 21 (15.11.2013): 4170. http://dx.doi.org/10.1182/blood.v122.21.4170.4170.
Pełny tekst źródłaTripathi, Malati, Ayushma Adhikari i Bibhushan Neupane. "Misoprostol Versus Oxytocin for Induction of Labour at Term and Post Term Pregnancy of Primigravida". Journal of Universal College of Medical Sciences 6, nr 2 (3.12.2018): 56–59. http://dx.doi.org/10.3126/jucms.v6i2.22497.
Pełny tekst źródłaOlupot-Olupot, Peter, Florence Aloroker, Ayub Mpoya, Hellen Mnjalla, George Passi, Margaret Nakuya, Kirsty Houston i in. "Gastroenteritis Rehydration Of children with Severe Acute Malnutrition (GASTROSAM): A Phase II Randomised Controlled trial: Trial Protocol". Wellcome Open Research 6 (23.06.2021): 160. http://dx.doi.org/10.12688/wellcomeopenres.16885.1.
Pełny tekst źródłaJiang, Jianhua, Yangang Ren, Yang Guan, Shengbo Eben Li, Yuming Yin, Dongjie Yu i Xiaoping Jin. "Integrated decision and control at multi-lane intersections with mixed traffic flow". Journal of Physics: Conference Series 2234, nr 1 (1.04.2022): 012015. http://dx.doi.org/10.1088/1742-6596/2234/1/012015.
Pełny tekst źródła