Neural network model for autonomous navigation of a water drone
DOI:
https://doi.org/10.20535/2786-8729.5.2024.315700Keywords:
water drone, machine learning, DQN, autonomous navigation, algorithmAbstract
Water drones have significant potential for use in environmental monitoring, search and rescue operations, and marine infrastructure inspection, but the specific conditions of the water environment make it difficult to implement stable autonomous navigation.
The object of research presented in this paper is the machine learning process for autonomous navigation of a water drone model in a simulated water environment. The purpose of the study is to implement a neural network model for autonomous navigation of a water drone using a reinforcement learning method that provides improved obstacle avoidance and adaptation to water currents.
To achieve this purpose, a new neural network model for autonomous drone navigation in the water environment based on the reinforcement learning method is proposed, which differs from the existing ones in that it uses an improved drone control algorithm that takes into account the speed and direction of the water current, which makes it possible to stabilize the process of generating neural network coefficients.
To ensure an effective learning process and optimization of the model, a simulation training environment was developed using the USVSim simulator, which contains various factors that interfere with the drone's movement, such as water current and the presence of other objects. The water drone, acting as an agent, gradually learns to choose the most effective actions to maximize positive rewards through trial and error, interacting with the environment and adapting to changing conditions. This process takes place through the use of a Deep Q-Network: the drone provides the value of its current state to a deep neural network; the neural network processes the data, predicts the value of the most effective action, and gives it to the agent. The current state of the drone is information in the form of a set of sensor readings measuring the distance to the nearest obstacles, drone’s heading and current distance to goal. The value of the effective action received from the neural network is converted into a command for the rudder that the drone can understand. The value of the drone's thruster power is calculated by separate formulas using trigonometric functions.
The results of the study showed that the use of the proposed model allows the drone to make decisions in a dynamic water environment when rapid adaptation to changes is required. The model successfully adapted its strategy based on feedback from the environment, so it can be concluded that the implemented model shows significant potential for further research and applications in the field of autonomous water drones, especially in changing and unpredictable environments.
References
G. Katsouras et al., “Use of Unmanned Surface Vehicles (USVs) in Water Chemistry Studies,” Sensors, vol. 24, no. 9, p. 2809, Apr. 2024. https://doi.org/10.3390/s24092809.
Z. Liu, Y. Zhang, X. Yu, and C. Yuan, “Unmanned surface vehicles: An overview of developments and challenges,” Annual Reviews in Control, vol. 41, pp. 71–93, 2016. https://doi.org/10.1016/j.arcontrol.2016.04.018.
A. Bhar and M. Sayadi, “On designing a configurable UAV autopilot for unmanned quadrotors,” Frontiers in Neurorobotics, vol. 18, May. 2024. https://doi.org/10.3389/fnbot.2024.1363366.
A. Kumar et al., “Enhanced Route navigation control system for turtlebot using human-assisted mobility and 3-D SLAM optimization”, Heliyon, vol. 10, p. e26828, Mar. 2024. https://doi.org/10.1016/j.heliyon.2024.e26828.
A. Marashian and A. Razminia, “Mobile robot’s path-planning and path-tracking in static and dynamic environments: Dynamic programming approach,”, Robotics and Autonomous Systems, vol. 172, p. 104592, Feb. 2024. https://doi.org/10.1016/j.robot.2023.104592.
A. Bhargava, Grokking Algorithms, 2nd ed., New York, NY, US: Manning Publications Co. LLC, 2024.
T. Xu, “Recent advances in Rapidly-exploring random tree: A review,” Heliyon, vol. 10, p. e32451, Jun. 2024. https://doi.org/10.1016/j.heliyon.2024.e32451.
R. S. Sutton and A. G. Barto, “Reinforcement Learning: An Introduction,” IEEE Transactions on Neural Networks, vol. 9, no. 5, p. 1054, Sep. 1998. https://doi.org/10.1109/tnn.1998.712192.
G. Wu, D. Li, H. Ding, D. Shi, and B. Han, “An overview of developments and challenges for unmanned surface vehicle autonomous berthing,” Complex & Intelligent Systems, vol. 10, pp. 981–1003, Aug. 2023. https://doi.org/10.1007/s40747-023-01196-z.
Y. Zhao, F. Han, D. Han, X. Peng, and W. Zhao, “Decision-making for the autonomous navigation of USVs based on deep reinforcement learning under IALA maritime buoyage system,” Ocean Engineering, vol. 266, p. 112557, Dec. 2022. https://doi.org/10.1016/j.oceaneng.2022.112557.
L. Li, D. Wu, Y. Huang, and Z. Yuan, “A path planning strategy unified with a COLREGS collision avoidance function based on deep reinforcement learning and artificial potential field,” Applied Ocean Research, vol. 113, p. 102759, Aug. 2021. https://doi.org/10.1016/j.apor.2021.102759.
K. Messaoudi, O. S. Oubbati, A. Rachedi, A. Lakas, T. Bendouma, and N. Chaib, “A survey of UAV-based data collection: Challenges, solutions and future perspectives,” Journal of Network and Computer Applications, vol. 216, p. 103670, Jul. 2023. https://doi.org/10.1016/j.jnca.2023.103670.
R. Song, Y. Liu, and R. Bucknall, “Smoothed A* algorithm for practical unmanned surface vehicle path planning,” Applied Ocean Research, vol. 83, pp. 9–20, Feb. 2019. https://doi.org/10.1016/j.apor.2018.12.001.
Hugging Face. “The Deep Q-Learning Algorithm”. Accessed: Nov. 15, 2024. [Online]. Available: https://huggingface.co/learn/deep-rl-course/unit3/deep-q-algorithm.
M. Roderick, J. MacGlashan, and S. Tellex, “Implementing the Deep Q-Network,” Nov. 2017, arXiv:1711.07478. https://doi.org/10.48550/arXiv.1711.07478.
M. Paravisi, D. H. Santos, V. Jorge, G. Heck, L. M. Gonçalves, and A. Amory, “Unmanned Surface Vehicle Simulator with Realistic Environmental Disturbances,” Sensors, vol. 19, no. 5, p. 1068, Mar. 2019. https://doi.org/10.3390/s19051068.
J. F. Steffensen, Interpolation, 2nd ed., Mineola, NY, US: Dover Publications, 2006.
F. N. Fritsch and R. E. Carlson, “Monotone Piecewise Cubic Interpolation,” SIAM Journal on Numerical Analysis, vol. 17, no. 2, pp. 238–246, Apr. 1980. https://doi.org/10.1137/0717021.
J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov, “Proximal Policy Optimization Algorithms,” Jul. 2017, arXiv:1707.06347. https://doi.org/10.48550/arXiv.1707.06347.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2024 Information, Computing and Intelligent systems
This work is licensed under a Creative Commons Attribution 4.0 International License.