Gotowa bibliografia na temat „Deep Learning and Perception for Grasping and Manipulation”
Utwórz poprawne odniesienie w stylach APA, MLA, Chicago, Harvard i wielu innych
Zobacz listy aktualnych artykułów, książek, rozpraw, streszczeń i innych źródeł naukowych na temat „Deep Learning and Perception for Grasping and Manipulation”.
Przycisk „Dodaj do bibliografii” jest dostępny obok każdej pracy w bibliografii. Użyj go – a my automatycznie utworzymy odniesienie bibliograficzne do wybranej pracy w stylu cytowania, którego potrzebujesz: APA, MLA, Harvard, Chicago, Vancouver itp.
Możesz również pobrać pełny tekst publikacji naukowej w formacie „.pdf” i przeczytać adnotację do pracy online, jeśli odpowiednie parametry są dostępne w metadanych.
Artykuły w czasopismach na temat "Deep Learning and Perception for Grasping and Manipulation"
Han, Dong, Hong Nie, Jinbao Chen, Meng Chen, Zhen Deng i Jianwei Zhang. "Multi-modal haptic image recognition based on deep learning". Sensor Review 38, nr 4 (17.09.2018): 486–93. http://dx.doi.org/10.1108/sr-08-2017-0160.
Pełny tekst źródłaValarezo Añazco, Edwin, Sara Guerrero, Patricio Rivera Lopez, Ji-Heon Oh, Ga-Hyeon Ryu i Tae-Seong Kim. "Deep Learning-Based Ensemble Approach for Autonomous Object Manipulation with an Anthropomorphic Soft Robot Hand". Electronics 13, nr 2 (17.01.2024): 379. http://dx.doi.org/10.3390/electronics13020379.
Pełny tekst źródłaWang, Cong, Qifeng Zhang, Qiyan Tian, Shuo Li, Xiaohui Wang, David Lane, Yvan Petillot i Sen Wang. "Learning Mobile Manipulation through Deep Reinforcement Learning". Sensors 20, nr 3 (10.02.2020): 939. http://dx.doi.org/10.3390/s20030939.
Pełny tekst źródłaZhao, Wenhui, Bin Xu i Xinzhong Wu. "Robot grasping system based on deep learning target detection". Journal of Physics: Conference Series 2450, nr 1 (1.03.2023): 012071. http://dx.doi.org/10.1088/1742-6596/2450/1/012071.
Pełny tekst źródłaZhou, Hongyu, Jinhui Xiao, Hanwen Kang, Xing Wang, Wesley Au i Chao Chen. "Learning-Based Slip Detection for Robotic Fruit Grasping and Manipulation under Leaf Interference". Sensors 22, nr 15 (22.07.2022): 5483. http://dx.doi.org/10.3390/s22155483.
Pełny tekst źródłaZhang, Ruihua, Xujun Chen, Zhengzhong Wan, Meng Wang i Xinqing Xiao. "Deep Learning-Based Oyster Packaging System". Applied Sciences 13, nr 24 (8.12.2023): 13105. http://dx.doi.org/10.3390/app132413105.
Pełny tekst źródłaLiu, Ning, Cangui Guo, Rongzhao Liang i Deping Li. "Collaborative Viewpoint Adjusting and Grasping via Deep Reinforcement Learning in Clutter Scenes". Machines 10, nr 12 (29.11.2022): 1135. http://dx.doi.org/10.3390/machines10121135.
Pełny tekst źródłaHan, Dong, Beni Mulyana, Vladimir Stankovic i Samuel Cheng. "A Survey on Deep Reinforcement Learning Algorithms for Robotic Manipulation". Sensors 23, nr 7 (5.04.2023): 3762. http://dx.doi.org/10.3390/s23073762.
Pełny tekst źródłaMohammed, Marwan Qaid, Lee Chung Kwek, Shing Chyi Chua, Abdulaziz Salamah Aljaloud, Arafat Al-Dhaqm, Zeyad Ghaleb Al-Mekhlafi i Badiea Abdulkarem Mohammed. "Deep Reinforcement Learning-Based Robotic Grasping in Clutter and Occlusion". Sustainability 13, nr 24 (10.12.2021): 13686. http://dx.doi.org/10.3390/su132413686.
Pełny tekst źródłaSayour, Malak H., Sharbel E. Kozhaya i Samer S. Saab. "Autonomous Robotic Manipulation: Real-Time, Deep-Learning Approach for Grasping of Unknown Objects". Journal of Robotics 2022 (30.06.2022): 1–14. http://dx.doi.org/10.1155/2022/2585656.
Pełny tekst źródłaRozprawy doktorskie na temat "Deep Learning and Perception for Grasping and Manipulation"
Zapata-Impata, Brayan S. "Robotic manipulation based on visual and tactile perception". Doctoral thesis, Universidad de Alicante, 2020. http://hdl.handle.net/10045/118217.
Pełny tekst źródłaThis doctoral thesis has been carried out with the support of the Spanish Ministry of Economy, Industry and Competitiveness through the grant BES-2016-078290.
Tahoun, Mohamed. "Object Shape Perception for Autonomous Dexterous Manipulation Based on Multi-Modal Learning Models". Electronic Thesis or Diss., Bourges, INSA Centre Val de Loire, 2021. http://www.theses.fr/2021ISAB0003.
Pełny tekst źródłaThis thesis proposes 3D object reconstruction methods based on multimodal deep learning strategies. The targeted applications concern robotic manipulation. First, the thesis proposes a 3D visual reconstruction method from a single view of the object obtained by an RGB-D sensor. Then, in order to improve the quality of 3D reconstruction of objects from a single view, a new method combining visual and tactile information has been proposed based on a learning reconstruction model. The proposed method has been validated on a visual-tactile dataset respecting the kinematic constraints of a robotic hand. The visual-tactile dataset respecting the kinematic properties of the multi-fingered robotic hand has been created in the framework of this PhD work. This dataset is unique in the literature and is also a contribution of the thesis. The validation results show that the tactile information can have an important contribution for the prediction of the complete shape of an object, especially the part that is not visible to the RGD-D sensor. They also show that the proposed model allows to obtain better results compared to those obtained with the best performing methods of the state of the art
Morrison, Douglas. "Robotic grasping in unstructured and dynamic environments". Thesis, Queensland University of Technology, 2021. https://eprints.qut.edu.au/207886/1/Douglas_Morrison_Thesis.pdf.
Pełny tekst źródłaCzęści książek na temat "Deep Learning and Perception for Grasping and Manipulation"
Blank, Andreas, Lukas Zikeli, Sebastian Reitelshöfer, Engin Karlidag i Jörg Franke. "Augmented Virtuality Input Demonstration Refinement Improving Hybrid Manipulation Learning for Bin Picking". W Lecture Notes in Mechanical Engineering, 332–41. Cham: Springer International Publishing, 2022. http://dx.doi.org/10.1007/978-3-031-18326-3_32.
Pełny tekst źródłaMehman Sefat, Amir, Saad Ahmad, Alexandre Angleraud, Esa Rahtu i Roel Pieters. "Robotic grasping in agile production". W Deep Learning for Robot Perception and Cognition, 407–33. Elsevier, 2022. http://dx.doi.org/10.1016/b978-0-32-385787-1.00021-x.
Pełny tekst źródłaKantor, George, i Francisco Yandun. "Advances in grasping techniques in agricultural robots". W Burleigh Dodds Series in Agricultural Science, 355–86. Burleigh Dodds Science Publishing, 2024. http://dx.doi.org/10.19103/as.2023.0124.09.
Pełny tekst źródła"Deep learning techniques for modelling human manipulation and its translation for autonomous robotic grasping with soft end-effe". W AI for Emerging Verticals: Human-robot computing, sensing and networking, 3–28. Institution of Engineering and Technology, 2020. http://dx.doi.org/10.1049/pbpc034e_ch1.
Pełny tekst źródłaStreszczenia konferencji na temat "Deep Learning and Perception for Grasping and Manipulation"
Chu, You-Rui, Haiyue Zhu i Zhiping Lin. "Intelligent 6-DoF Robotic Grasping and Manipulation System Using Deep Learning". W International Conference of Asian Society for Precision Engineering and Nanotechnology. Singapore: Research Publishing Services, 2022. http://dx.doi.org/10.3850/978-981-18-6021-8_or-02-0217.html.
Pełny tekst źródłaZhang, Chi, i Yingzhao Zhu. "A review of robot grasping tactile perception based on deep learning". W Third International Conference on Control and Intelligent Robotics (ICCIR 2023), redaktorzy Kechao Wang i M. Vijayalakshmi. SPIE, 2023. http://dx.doi.org/10.1117/12.3011588.
Pełny tekst źródłaPavlichenko, Dmytro, i Sven Behnke. "Deep Reinforcement Learning of Dexterous Pre-Grasp Manipulation for Human-Like Functional Categorical Grasping". W 2023 IEEE 19th International Conference on Automation Science and Engineering (CASE). IEEE, 2023. http://dx.doi.org/10.1109/case56687.2023.10260385.
Pełny tekst źródłaFang, Jianhao, Weifei Hu, Chuxuan Wang, Zhenyu Liu i Jianrong Tan. "Deep Reinforcement Learning Enhanced Convolutional Neural Networks for Robotic Grasping". W ASME 2021 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference. American Society of Mechanical Engineers, 2021. http://dx.doi.org/10.1115/detc2021-67225.
Pełny tekst źródłaLu, Jingpei, Ambareesh Jayakumari, Florian Richter, Yang Li i Michael C. Yip. "SuPer Deep: A Surgical Perception Framework for Robotic Tissue Manipulation using Deep Learning for Feature Extraction". W 2021 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2021. http://dx.doi.org/10.1109/icra48506.2021.9561249.
Pełny tekst źródłaRakhimkul, Sanzhar, Anton Kim, Askarbek Pazylbekov i Almas Shintemirov. "Autonomous Object Detection and Grasping Using Deep Learning for Design of an Intelligent Assistive Robot Manipulation System". W 2019 IEEE International Conference on Systems, Man and Cybernetics (SMC). IEEE, 2019. http://dx.doi.org/10.1109/smc.2019.8914465.
Pełny tekst źródłaImran, Alishba, William Escobar i Fred Barez. "Design of an Affordable Prosthetic Arm Equipped With Deep Learning Vision-Based Manipulation". W ASME 2021 International Mechanical Engineering Congress and Exposition. American Society of Mechanical Engineers, 2021. http://dx.doi.org/10.1115/imece2021-68714.
Pełny tekst źródłaChen, Zhu, Xiao Liang i Minghui Zheng. "Including Image-Based Perception in Disturbance Observer for Warehouse Drones". W ASME 2020 Dynamic Systems and Control Conference. American Society of Mechanical Engineers, 2020. http://dx.doi.org/10.1115/dscc2020-3284.
Pełny tekst źródła