Journal: Journal of Machine Learning and Deep Learning (JMLDL), Volume:1, Issue:1, Pages: 20-31 Download pdf
Authors: Mohammad Dehghani Tezerjani, Mohammad Khoshnazar, Mohammadhamed Tangestanizadeh,
Arman Kiani, Qing Yang
Date: 12-2024
Abstract:The emergence of mobile robotics, particularly in the automotive industry, introduces a promising era of enriched user experiences and adept handling of complex navigation challenges. The realization of these advancements necessitates a focused technological effort and the successful execution of numerous intricate tasks, particularly in the critical domain of Simultaneous Localization and Mapping (SLAM). Various Artificial Intelligence (AI) methodologies, such as deep learning and Reinforcement Learning (RL), present viable solutions to address the challenges in SLAM. This study specifically explores the application of RL in the context of SLAM. By enabling the agent (the robot) to iteratively interact with and receive feedback from its environment, RL facilitates the acquisition of navigation and mapping skills, thereby enhancing the robot's decision-making capabilities. This approach offers several advantages, including improved navigation proficiency, increased resilience, reduced dependence on sensor precision, and refinement of the decision-making process. The findings of this study, which provides an overview of RL's utilization in SLAM, reveal significant advancements in the field. The investigation also highlights the evolution and innovative integration of these techniques.
Keywords: Simultaneous Localization and Mapping, Reinforcement Learning, Path Planning, Loop Closure Detection, Active SLAM.
References:
[1] N. Sharma, J. K. Pandey, and S. Mondal, "A review of mobile robots: Applications and future prospect," International Journal of Precision Engineering and Manufacturing, vol. 24, no. 9, pp. 1695–1706, Sep. 2023, doi: 10.1007/s12541-023-00876-7.
[2] K. Othman, "Public acceptance and perception of autonomous vehicles: A comprehensive review," AI and Ethics, vol. 1, no. 3, pp. 355–387, Aug. 2021, doi: 10.1007/s43681-021-00041-8.
[3] G. Bresson, Z. Alsayed, L. Yu, and S. Glaser, "Simultaneous localization and mapping: A survey of current trends in autonomous driving," IEEE Transactions on Intelligent Vehicles, vol. 2, no. 3, pp. 194–220, 2017.
[4] M. F. Ahmed, K. Masood, V. Fremont, and I. Fantoni, "Active SLAM: A review on last decade," Sensors, vol. 23, no. 19, p. 8097, 2023. [Online]. Available: https://www.mdpi.com/1424-8220/23/19/8097.
[5] J. Ren and D. Xia, "SLAM in autonomous driving," in Autonomous Driving Algorithms and Its IC Design. Singapore: Springer Nature Singapore, 2023, pp. 127–152.
[6] J. A. Placed et al., "A survey on active simultaneous localization and mapping: State of the art and new frontiers," IEEE Transactions on Robotics, 2023.
[7] X. Wu, G. Wang, and N. Shen, "Research on obstacle avoidance optimization and path planning of autonomous vehicles based on attention mechanism combined with multimodal information decision-making thoughts of robots," Frontiers in Neurorobotics, vol. 17, p. 1269447, 2023, doi: 10.3389/fnbot.2023.1269447.
[8] H. Taheri and Z. C. Xia, "SLAM; definition and evolution," Engineering Applications of Artificial Intelligence, vol. 97, p. 104032, Jan. 2021, doi: https://doi.org/10.1016/j.engappai.2020.104032.
[9] R. Zeng, Y. Wen, W. Zhao, and Y.-J. Liu, "View planning in robot active vision: A survey of systems, algorithms, and applications," Computational Visual Media, vol. 6, pp. 225–245, 2020.
[10] M. Diginsa, N. Shafie, and N. Yusuf, "Review: Issues and challenges of simultaneous localization and mapping (SLAM) technology in autonomous robot," International Journal of Innovative Computing, vol. 13, pp. 59–63, Nov. 2023, doi: 10.11113/ijic.v13n2.408.
[11] R. Eyvazpour, M. Shoaran, and G. Karimian, "Hardware implementation of SLAM algorithms: A survey on implementation approaches and platforms," Artificial Intelligence Review, vol. 56, no. 7, pp. 6187–6239, Jul. 2023, doi: 10.1007/s10462-022-10310-5.
[12] S. Li, Z. Li, X. Liu, C. Shan, Y. Zhao, and H. Cheng, "Research on map-SLAM fusion localization algorithm for unmanned vehicle," Applied Sciences, vol. 12, no. 17, p. 8670, 2022. [Online]. Available: https://www.mdpi.com/2076-3417/12/17/8670.
[13] T. Liu, C. Xu, Y. Qiao, C. Jiang, and J. Yu, "Particle filter SLAM for vehicle localization," arXiv preprint, arXiv:2402.07429, 2024.
[14] C. Fan, Z. Li, W. Ding, H. Zhou, and K. Qian, Integrating Artificial Intelligence with SLAM Technology for Robotic Navigation and Localization in Unknown Environments, 2024.
[15] N. Botteghi, B. Kallfelz Sirmacek, R. Schulte, M. Poel, and C. Brune, Reinforcement Learning Helps SLAM: Learning to Build Maps, 2020, pp. 329–335.
[16] A. Hamza, "Deep reinforcement learning for mapless mobile robot navigation," 2022.
[17] D. Nister and H. Stewenius, "Scalable recognition with a vocabulary tree," in 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06), 2006, vol. 2, pp. 2161–2168.
[18] S. Saeedi, L. Paull, M. Trentini, and H. Li, "Multiple robot simultaneous localization and mapping," in 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2011, pp. 853–858.
[19] G. Grisetti, R. Kümmerle, C. Stachniss, and W. Burgard, "A tutorial on graph-based SLAM," IEEE Intelligent Transportation Systems Magazine, vol. 2, no. 4, pp. 31–43, 2010.
[20] S. Frintrop, P. Jensfelt, and H. Christensen, Simultaneous Robot Localization and Mapping Based on a Visual Attention System, 2007, pp. 417–430.
[21] S. Thrun, "Simultaneous localization and mapping," in Robotics and Cognitive Approaches to Spatial Mapping. Springer, 2008, pp. 13–41.
[22] S. Zheng, J. Wang, C. Rizos, W. Ding, and A. El-Mowafy, "Simultaneous localization and mapping (SLAM) for autonomous driving: Concept and analysis," Remote Sensing, vol. 15, no. 4, p. 1156, 2023. [Online]. Available: https://www.mdpi.com/2072-4292/15/4/1156.
[23] H. Carrillo, I. Reid, and J. A. Castellanos, "On the comparison of uncertainty criteria for active SLAM," in 2012 IEEE International Conference on Robotics and Automation. IEEE, 2012, pp. 2080–2087.
[24] W. Burgard, D. Fox, and S. Thrun, "Active mobile robot localization," in IJCAI. Citeseer, 1997, pp. 1346–1352.
[25] A. A. Makarenko, S. B. Williams, F. Bourgault, and H. F. Durrant-Whyte, "An experiment in integrated exploration," in IEEE/RSJ International Conference on Intelligent Robots and Systems, vol. 1. IEEE, 2002, pp. 534–539.
[26] J. Lin, J. Peng, Z. Hu, X. Xie, and R. Peng, "ORB-SLAM, IMU and wheel odometry fusion for indoor mobile robot localization and navigation," Academic Journal of Computer and Information Science, vol. 3, 2020.
[27] A. Kumar, Maneesha, and P. K. Pandey, "Advances in simultaneous localization and mapping (SLAM) for autonomous mobile robot navigation," in Proceedings of International Joint Conference on Advances in Computational Intelligence. Springer Nature Singapore, Singapore, 2024, pp. 481–493.
[28] E. Reitbauer, C. Schmied, F. Theurl, and M. Wieser, "LIWO-SLAM: A LiDAR, IMU, and wheel odometry simultaneous localization and mapping system for GNSS-denied environments based on factor graph optimization," in Proceedings of the 36th International Technical Meeting of the Satellite Division of The Institute of Navigation (ION GNSS+ 2023), 2023, pp. 1669–1683.
[29] S. Nam, C. Woo, S. Kang, T. A. Nguyen, and D. Min, "SLAM-DRLnav: A SLAM-enhanced deep reinforcement learning navigation framework for indoor self-driving," in 2023 International Conference on Mechatronics, Control and Robotics (ICMCR), Feb. 2023, pp. 44–48, doi: 10.1109/ICMCR56776.2023.10181042.
[30] G. Dulac-Arnold et al., "Challenges of real-world reinforcement learning: Definitions, benchmarks and analysis," Machine Learning, vol. 110, no. 9, pp. 2419–2468, Sep. 2021, doi: 10.1007/s10994-021-05961-4.
[31] C. Debeunne and D. Vivet, "A review of visual-LiDAR fusion based simultaneous localization and mapping," Sensors, vol. 20, no. 7, p. 2068, 2020.
[32] J. Tang et al., "LiDAR scan matching aided inertial navigation system in GNSS-denied environments," Sensors, vol. 15, no. 7, pp. 16710–16728, 2015.
[33] Y. Yang, D. Tang, D. Wang, W. Song, J. Wang, and M. Fu, "Multi-camera visual SLAM for off-road navigation," Robotics and Autonomous Systems, vol. 128, p. 103505, 2020.
[34] X. Lang et al., "Gaussian-LIC: Photo-realistic LiDAR-Inertial-Camera SLAM with 3D Gaussian splatting," arXiv preprint, arXiv:2404.06926, 2024.
[35] L. Chang, X. Niu, and T. Liu, "GNSS/IMU/ODO/LiDAR-SLAM integrated navigation system using IMU/ODO pre-integration," Sensors, vol. 20, p. 4702, Aug. 2020, doi: 10.3390/s20174702.
[36] N. Abdelaziz and A. El-Rabbany, "INS/LIDAR/Stereo SLAM integration for precision navigation in GNSS-denied environments," Sensors, vol. 23, no. 17, p. 7424, 2023. [Online]. Available: https://www.mdpi.com/1424-8220/23/17/7424.
[37] A. Woo, B. Fidan, and W. W. Melek, "Localization for autonomous driving," in Handbook of Position Location: Theory, Practice, and Advances, 2nd ed., 2018, pp. 1051–1087.
[38] M. S. A. Khan, D. Hussain, K. Naveed, U. S. Khan, I. Q. Mundial, and A. B. Aqeel, "Investigation of widely used SLAM sensors using analytical hierarchy process," Journal of Sensors, vol. 2022, pp. 1–15, 2022.
[39] Z. Hong, Y. Petillot, A. Wallace, and S. Wang, "Radar SLAM: A robust SLAM system for all weather conditions," arXiv preprint, arXiv:2104.05347, 2021.
[40] R. Huang, K. Zhu, S. Chen, T. Xiao, M. Yang, and N. Zheng, "A high-precision and robust odometry based on sparse MMW radar data and a large-range and long-distance radar positioning data set," in 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), 2021, pp. 98–105.
[41] J. Zhu, H. Li, and T. Zhang, "Camera, LiDAR, and IMU based multi-sensor fusion SLAM: A survey," Tsinghua Science and Technology, vol. 29, no. 2, pp. 415–429, 2023.
[42] Y.-E. Lu, S. Tsai, M.-L. Tsai, and K.-W. Chiang, "A low-cost visual radar-based odometry framework with mmWave radar and monocular camera," The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, vol. 43, pp. 235–240, 2022.
[43] M. Quan, S. Piao, M. Tan, and S.-S. Huang, "Tightly-coupled monocular visual-odometric SLAM using wheels and a MEMS gyroscope," IEEE Access, vol. 7, pp. 97374–97389, 2019.
[44] K. Yousif, A. Bab-Hadiashar, and R. Hoseinnezhad, "An overview to visual odometry and visual SLAM: Applications to mobile robotics," Intelligent Industrial Systems, vol. 1, no. 4, pp. 289–311, Dec. 2015, doi: 10.1007/s40903-015-0032-7.
[45] J. Mahmoud and A. Penkovskiy, Dynamic Environments and Robust SLAM: Optimizing Sensor Fusion and Semantics for Wheeled Robots, 2023, pp. 185–191.
[46] I. H. Sarker, "Machine learning: Algorithms, real-world applications and research directions," SN Computer Science, vol. 2, no. 3, p. 160, 2021.
[47] R. S. Sutton and A. G. Barto, Reinforcement Learning: An Introduction. MIT Press, 2018.
[48] R. Singh, J. Ren, and X. Lin, "A review of deep reinforcement learning algorithms for mobile robot path planning," Vehicles, vol. 5, no. 4, pp. 1423–1451, 2023. [Online]. Available: https://www.mdpi.com/2624-8921/5/4/78.
[49] K. Arulkumaran, M. P. Deisenroth, M. Brundage, and A. A. Bharath, "A brief survey of deep reinforcement learning," arXiv preprint, arXiv:1708.05866, 2017.
[50] M. Sewak, S. K. Sahay, and H. Rathore, "Policy-approximation based deep reinforcement learning techniques: An overview," in Information and Communication Technology for Competitive Strategies (ICTCS 2020): ICT Applications and Social Interfaces, 2022, pp. 493–507.
[51] H. Byeon, "Advances in value-based, policy-based, and deep learning-based reinforcement learning," International Journal of Advanced Computer Science and Applications, vol. 14, no. 8, 2023.
[52] J. Clifton and E. Laber, "Q-learning: Theory and applications," Annual Review of Statistics and Its Application, vol. 7, no. 1, pp. 279–301, 2020.
[53] Y. Huang, "Deep Q-networks," in Deep Reinforcement Learning: Fundamentals, Research and Applications, 2020, pp. 135–160.
[54] A. Momenikorbekandi and M. Abbod, "Intelligent scheduling based on reinforcement learning approaches: Applying advanced Q-learning and state–action–reward–state–action reinforcement learning models for the optimisation of job shop scheduling problems," Electronics, vol. 12, no. 23, p. 4752, 2023.
[55] S. Zhao and S.-H. Hwang, "Exploration-and exploitation-driven deep deterministic policy gradient for active SLAM in unknown indoor environments," Electronics, vol. 13, no. 5, p. 999, 2024.
[56] R. J. Williams, "Simple statistical gradient-following algorithms for connectionist reinforcement learning," Machine Learning, vol. 8, pp. 229–256, 1992.
[57] J. Schulman, "Trust region policy optimization," arXiv preprint, arXiv:1502.05477, 2015.
[58] V. T. Potter, Exploring Deep Reinforcement Learning Techniques for Autonomous Navigation, 2022.
[59] J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov, "Proximal policy optimization algorithms," arXiv preprint, arXiv:1707.06347, 2017.
[60] D. Joel, Y. Niv, and E. Ruppin, "Actor–critic models of the basal ganglia: New anatomical and computational perspectives," Neural Networks, vol. 15, no. 4–6, pp. 535–547, 2002.
[61] I. Grondman, M. Vaandrager, L. Busoniu, R. Babuska, and E. Schuitema, "Efficient model learning methods for actor–critic control," IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), vol. 42, no. 3, pp. 591–602, 2011.
[62] X. Wang et al., "Deep reinforcement learning: A survey," IEEE Transactions on Neural Networks and Learning Systems, 2022.
[63] A. Zai and B. Brown, Deep Reinforcement Learning in Action. Manning Publications, 2020.
[64] X. Lei, Z. Zhang, and P. Dong, "Dynamic path planning of unknown environment based on deep reinforcement learning," Journal of Robotics, vol. 2018, 2018.
[65] M.-F. R. Lee and S. H. Yusuf, "Mobile robot navigation using deep reinforcement learning," Processes, vol. 10, no. 12, p. 2748, 2022. [Online]. Available: https://www.mdpi.com/2227-9717/10/12/2748.
[66] C.-C. Wong, H.-M. Feng, and K.-L. Kuo, "Multi-sensor fusion simultaneous localization mapping based on deep reinforcement learning and multi-model adaptive estimation," Sensors, vol. 24, no. 1, p. 48, 2024. [Online]. Available: https://www.mdpi.com/1424-8220/24/1/48.
[67] W. Su, T. Huang, F. Liu, and H. Wang, "ASLAM-FD: A multi-sensor adaptive collaborative fusion SLAM framework based on degradation detection and deep reinforcement learning," Measurement Science and Technology, vol. 35, no. 12, p. 126312, Sep. 2024, doi: 10.1088/1361-6501/ad7bdd.
[68] X. Liu, S. Wen, Y. Hu, F. Han, H. Zhang, and H. R. Karimi, "An active SLAM with multi-sensor fusion for snake robots based on deep reinforcement learning," Mechatronics, vol. 103, p. 103248, Nov. 2024, doi: https://doi.org/10.1016/j.mechatronics.2024.103248.
[69] Q. Wang, W. Zhuang, L. Wang, and F. Ju, "Lane keeping assist for an autonomous vehicle based on deep reinforcement learning," SAE Technical Paper, 0148-7191, 2020.
[70] C. Arvind and J. Senthilnath, "Autonomous vehicle for obstacle detection and avoidance using reinforcement learning," in Soft Computing for Problem Solving: SocProS 2018, Volume 1, Springer, 2020, pp. 55–66.
[71] K. Rezaee, P. Yadmellat, and S. Chamorro, "Motion planning for autonomous vehicles in the presence of uncertainty using reinforcement learning," in 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), IEEE, 2021, pp. 3506–3511.
[72] C. Tang, B. Abbatematteo, J. Hu, R. Chandra, R. Martín-Martín, and P. Stone, "Deep reinforcement learning for robotics: A survey of real-world successes," arXiv preprint, arXiv:2408.03539, 2024.
[73] Y. Song, A. Romero, M. Müller, V. Koltun, and D. Scaramuzza, "Reaching the limit in autonomous racing: Optimal control versus reinforcement learning," Science Robotics, vol. 8, no. 82, p. eadg1462, 2023.
[74] M. Baziyad, M. Saad, R. Fareh, T. Rabie, and I. Kamel, "Addressing real-time demands for robotic path planning systems: A routing protocol approach," IEEE Access, vol. 9, pp. 38132–38143, 2021.
[75] S.-Y. Chen, Q.-F. He, and C.-F. Lai, "Deep reinforcement learning-based robot exploration for constructing map of unknown environment," Information Systems Frontiers, vol. 26, no. 1, pp. 63–74, 2024.
[76] W. Ruiqi, "Research on robot path planning based on reinforcement learning," arXiv preprint, arXiv:2404.14077, 2024.
[77] S.-Y. Chen, Q.-F. He, and C.-F. Lai, "Deep reinforcement learning-based robot exploration for constructing map of unknown environment," Information Systems Frontiers, vol. 26, Nov. 2021, doi: 10.1007/s10796-021-10218-5.
[78] H. Li, Q. Zhang, and D. Zhao, "Deep reinforcement learning-based automatic exploration for navigation in unknown environment," IEEE Transactions on Neural Networks and Learning Systems, vol. 31, no. 6, pp. 2064–2076, 2020, doi: 10.1109/TNNLS.2019.2927869.
[79] A. Iqbal, "Obstacle detection and track detection in autonomous cars," 2020.
[80] S. Wen, Y. Zhao, X. Yuan, Z. Wang, D. Zhang, and L. Manfredi, "Path planning for active SLAM based on deep reinforcement learning under unknown environments," Intelligent Service Robotics, vol. 13, pp. 263–272, 2020.
[81] S. Nam, C. Woo, S. Kang, T. A. Nguyen, and D. Min, "iNAV-drlSLAM: An improved indoor self-driving framework for mobile robots using deep reinforcement learning integrated with SLAM," in 2023 15th International Conference on Advanced Computational Intelligence (ICACI), May 2023, pp. 1–8, doi: 10.1109/ICACI58115.2023.10146173.
[82] A. R. Fayjie, S. Hossain, D. Oualid, and D. J. Lee, "Driverless car: Autonomous driving using deep reinforcement learning in urban environment," in 2018 15th International Conference on Ubiquitous Robots (UR), Jun. 2018, pp. 896–901, doi: 10.1109/URAI.2018.8441797.
[83] S. Arshad and G.-W. Kim, "Role of deep learning in loop closure detection for visual and lidar SLAM: A survey," Sensors, vol. 21, no. 4, p. 1243, 2021.
[84] K. A. Tsintotas, L. Bampis, and A. Gasteratos, "The revisiting problem in simultaneous localization and mapping: A survey on visual loop closure detection," IEEE Transactions on Intelligent Transportation Systems, vol. 23, no. 11, pp. 19929–19953, 2022.
[85] A. Charroud, K. El Moutaouakil, V. Palade, A. Yahyaouy, U. Onyekpe, and E. U. Eyo, "Localization and mapping for self-driving vehicles: A survey," Machines, vol. 12, no. 2, p. 118, 2024. [Online]. Available: https://www.mdpi.com/2075-1702/12/2/118.
[86] A. Iqbal, R. Thapa, and N. R. Gans, "Deep reinforcement learning based loop closure detection," Journal of Intelligent & Robotic Systems, vol. 106, no. 2, p. 51, 2022.
[87] A. Iqbal, Classified Object Localization in SLAM and Loop Closure Through Reinforcement Learning. The University of Texas at Dallas, 2019.
[88] J. A. Placed and J. A. Castellanos, "A deep reinforcement learning approach for active SLAM," Applied Sciences, vol. 10, no. 23, p. 8386, 2020. [Online]. Available: https://www.mdpi.com/2076-3417/10/23/8386.
[89] Y. Chen, S. Huang, and R. Fitch, "Active SLAM for mobile robots with area coverage and obstacle avoidance," IEEE/ASME Transactions on Mechatronics, vol. 25, no. 3, pp. 1182–1192, 2020.
[90] N. Palomeras, M. Carreras, and J. Andrade-Cetto, "Active SLAM for autonomous underwater exploration," Remote Sensing, vol. 11, no. 23, p. 2827, 2019.
[91] B. Fang, X. Chen, Z. Pan, and X. Di, "SLAMuZero: Plan and learn to map for joint SLAM and navigation," in 34th International Conference on Automated Planning and Scheduling.
[92] Z. Pei, S. Piao, M. Quan, M. Z. Qadir, and G. Li, "Active collaboration in relative observation for multi-agent visual simultaneous localization and mapping based on deep Q network," International Journal of Advanced Robotic Systems, vol. 17, no. 2, p. 1729881420920216, 2020.
[93] M. Alcalde, M. Ferreira, P. González, F. Andrade, and G. Tejera, "DA-SLAM: Deep active SLAM based on deep reinforcement learning," in 2022 Latin American Robotics Symposium (LARS), 2022 Brazilian Symposium on Robotics (SBR), and 2022 Workshop on Robotics in Education (WRE). IEEE, 2022, pp. 282–287.
[94] R. Zhao, Y. Li, Y. Fan, F. Gao, M. Tsukada, and Z. Gao, "A survey on recent advancements in autonomous driving using deep reinforcement learning: Applications, challenges, and solutions," IEEE Transactions on Intelligent Transportation Systems, 2024.
[95] A. Tourani, H. Bavle, J. L. Sanchez-Lopez, and H. Voos, "Visual SLAM: What are the current trends and what to expect?," Sensors, vol. 22, no. 23, p. 9297, 2022. [Online]. Available: https://www.mdpi.com/1424-8220/22/23/9297.
[96] M. N. Favorskaya, "Deep learning for visual SLAM: The state-of-the-art and future trends," Electronics, vol. 12, no. 9, p. 2006, 2023. [Online]. Available: https://www.mdpi.com/2079-9292/12/9/2006.
[97] H. Bavle, J. L. Sanchez-Lopez, C. Cimarelli, A. Tourani, and H. Voos, "From SLAM to situational awareness: Challenges and survey," Sensors, vol. 23, no. 10, p. 4849, 2023.
[98] C. Tessler, T. Zahavy, D. Cohen, D. J. Mankowitz, and S. Mannor, "Action assembly: Sparse imitation learning for text-based games with combinatorial action spaces," arXiv preprint, arXiv:1905.09700, 2019.
[99] A. Kudriashov, T. Buratowski, M. Giergiel, and P. Małka, SLAM Techniques Application for Mobile Robot in Rough Terrain. Springer, 2020.
[100] J. Buckman, D. Hafner, G. Tucker, E. Brevdo, and H. Lee, "Sample-efficient reinforcement learning with stochastic ensemble value expansion," Advances in Neural Information Processing Systems, vol. 31, 2018.
[101] C. Pang, L. Zhou, and X. Huang, "A low-cost 3D SLAM system integration of autonomous exploration based on fast-ICP enhanced LiDAR-inertial odometry," Remote Sensing, vol. 16, no. 11, p. 1979, 2024.
[102] T. Hester and P. Stone, "Texplore: Real-time sample-efficient reinforcement learning for robots," Machine Learning, vol. 90, pp. 385–429, 2013.
[103] N. Botteghi, B. Sirmacek, R. Schulte, M. Poel, and C. Brune, "Reinforcement learning helps SLAM: Learning to build maps," The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, vol. 43, pp. 329–335,