Energy-efficient Navigation in Unknown Static Flow Fields using Reinforcement Learning

Authors

Keywords:

Autonomous Agents, Reinforcement Learning, Intelligent Transportation

Abstract

Energy efficiency is becoming increasingly critical in the Maritime Industry due to expectations of cost reduction, environmental sustainability, regulatory compliance, trade volume growth, and safety enhancement. In this context, Reinforcement Learning (RL) methods have gained significant attention for solving control and optimization problems. Recent advancements have demonstrated effectiveness in diverse domains, including logistics, scheduling, and resource allocation. Hence, in this paper, we address the problem of energy-efficient navigation of vessels in unknown static flow fields using RL. We propose a reward function that minimizes control effort under flow influence and evaluate two tabular and two function-approximation methods across different scenarios with the state-of-the-art literature. Finally, we provide a zero-shot sim-to-sim analysis to evaluate the impact of environmental uncertainty on our proposed method.

Downloads

Download data is not yet available.

Author Biographies

Armando Alves Neto, Universidade Federal de Minas Gerais

Armando Alves Neto received the B.S.E. degree in Automation and Control Engineering from the Universidade Federal de Minas Gerais in 2006, and S.M. and Ph.D. degrees in Computer Science from UFMG in 2008 and 2012, respectively. He is an Assistant Professor at the Department of Electronic Engineering at UFMG. Research interests include real-time motion planning, multi-agent control, robust control, and collision avoidance strategies.

Victor Costa da Silva Campos, Universidade Federal de Minas Gerais

Victor Costa da Silva Campos is a Professor at the Dept. of Electronics Engineering at Universidade Federal de Minas Gerais (UFMG) since 2018. He received a B.Eng. degree in Control and Automation Engineering from UFMG in 2009, and M.Sc. and Ph.D. degrees in Electrical Engineering also from UFMG in 2011 and 2015, respectively. His research interests include robust and adaptive control, Takagi-Sugeno fuzzy systems, Computational Intelligence applied to control systems, and motion planning.

Douglas Guimarães Macharet, Universidade Federal de Minas Gerais

Douglas G. Macharet is an Associate Professor in the Dept. of Computer Science at the Universidade Federal de Minas Gerais (UFMG). He holds an M.Sc. and Ph.D. in Computer Science from UFMG, obtained in 2009 and 2013, respectively. He is with the Computer Vision and Robotics Laboratory (VeRLab), specializing in mobile robotics, with a focus on motion planning, navigation, multi-robot systems, swarm robotics, and human-robot interaction.

References

N. Gu, D. Wang, Z. Peng, J. Wang, and Q.-L. Han, “Advances in line-of-sight guidance for path following of autonomous marine vehicles: An overview,” IEEE Trans. on Systems, Man, and Cybernetics: Systems, vol. 53, no. 1, pp. 12–28, 2023. https://doi.org/10.1109/TSMC.2022.3162862.

S. M. LaValle, Planning Algorithms. New York, NY, USA: Cambridge University Press, 2006. https://doi.org/10.1017/CBO9780511546877.

K. Mallory, M. A. Hsieh, E. Forgoston, and I. B. Schwartz, “Distributed allocation of mobile sensing swarms in gyre flows,” Nonlin. Processes Geophys., vol. 20, no. 5, pp. 657–668, 2013. https://doi.org/10.5194/npg-20-657-2013.

C. R. Heckman, I. B. Schwartz, and M. A. Hsieh, “Toward efficient navigation in uncertain gyre-like flows,” The Int. J. of Robotics Research, vol. 34, no. 13, pp. 1590–1603, 2015. https://doi.org/10.1177/0278364915585396.

A. Alvarez, A. Caiti, and R. Onken, “Evolutionary path planning for autonomous underwater vehicles in a variable ocean,” IEEE J. of Oceanic Engineering, vol. 29, no. 2, pp. 418–429, 2004. https://doi.org/10.1109/JOE.2004.827837.

T. Lolla, P. F. Lermusiaux, M. P. Ueckermann, and P. J. Haley, “Time-optimal path planning in dynamic flows using level set equations: theory and schemes,” Ocean Dynamics, vol. 64, pp. 1373–1397, 2014. https://doi.org/10.1007/s10236-014-0757-y.

D. Kularatne, S. Bhattacharya, and M. A. Hsieh, “Going with the flow: a graph based approach to optimal path planning in general flows,” Autonomous Robots, vol. 42, pp. 1369–1387, 2018. https://doi.org/10.1007/s10514-018-9741-6.

M. K. Nutalapati, S. Joshi, and K. Rajawat, “Online utility-optimal trajectory design for time-varying ocean environments,” in Int. Conf. on Robotics and Automation (ICRA), pp. 6853–6859, IEEE, 2019. https://doi.org/10.1109/ICRA.2019.8794365.

H. R. Karimi and Y. Lu, “Guidance and control methodologies for marine vehicles: A survey,” Control Engineering Practice, vol. 111, p. 104785, 2021. https://doi.org/10.1016/j.conengprac.2021.104785.

B. Garau, A. Alvarez, and G. Oliver, “Path Planning of Autonomous Underwater Vehicles in Current Fields with Complex Spatial Variability: an A* Approach,” IEEE Int. Conf. on Robotics and Automation (ICRA), pp. 194–198, 2005. https://doi.org/10.1109/ROBOT.2005.1570118.

T.-B. Koay and M. Chitre, “Energy-efficient path planning for fully propelled auvs in congested coastal waters,” in MTS/IEEE OCEANS, pp. 1–9, 2013. https://doi.org/10.1109/OCEANS-Bergen.2013.6608168.

M. Otte, W. Silva, and E. Frew, “Any-time path-planning: Time-varying wind field + moving obstacles,” in IEEE Int. Conf. on Robotics and Automation, pp. 2575–2582, 2016. https://doi.org/10.1109/ICRA.2016.7487414.

K. C. To, K. M. B. Lee, C. Yoo, S. Anstee, and R. Fitch, “Streamlines for motion planning in underwater currents,” in Int. Conf. on Robotics and Automation, pp. 4619–4625, IEEE, 2019. https://doi.org/10.1109/ICRA.2019.8793567.

D. Kularatne, H. Hajieghrary, and M. A. Hsieh, “Optimal path planning in time-varying flows with forecasting uncertainties,” in IEEE Int. Conf. on Robotics and Automation, pp. 4857–4864, IEEE, 2018. https://doi.org/10.1109/ICRA.2018.8460221.

C. Yoo, J. J. H. Lee, S. Anstee, and R. Fitch, “Path planning in uncertain ocean currents using ensemble forecasts,” in IEEE Int. Conf. on Robotics and Automation, pp. 8323–8329, IEEE, 2021. https://doi.org/10.1109/ICRA48506.2021.9561626.

V. Mnih, K. Kavukcuoglu, D. Silver, A. Rusu, J. Veness, M. Bellemare, A. Graves, M. Riedmiller, A. Fidjeland, G. Ostrovski, S. Petersen, C. Beattie, A. Sadik, I. Antonoglou, H. King, D. Kumaran, D. Wierstra, S. Legg, and D. Hassabis, “Human-level control through Deep Reinforcement Learning,” Nature, vol. 518, no. 7540, pp. 529–533, 2015. https://doi.org/10.1038/nature14236.

Y. Qiao, J. Yin, W. Wang, F. Duarte, J. Yang, and C. Ratti, “Survey of Deep Learning for Autonomous Surface Vehicles in Marine Environments,” IEEE Trans. on Intelligent Transportation Systems, vol. 24, no. 4, pp. 3678–3701, 2023. https://doi.org/10.1109/TITS.2023.3235911.

B. Yoo and J. Kim, “Path optimization for marine vehicles in ocean currents using reinforcement learning,” J. of Marine Science and Technology, vol. 21, pp. 334–343, Jun 2016. https://doi.org/10.1007/s00773-015-0355-9.

W. Lan, X. Jin, X. Chang, T. Wang, H. Zhou, W. Tian, and L. Zhou, “Path planning for underwater gliders in time-varying ocean current using deep reinforcement learning,” Ocean Engineering, vol. 262, p. 112226, 2022. https://doi.org/10.1016/j.oceaneng.2022.112226.

Z. Chu, F. Wang, T. Lei, and C. Luo, “Path Planning Based on Deep Reinforcement Learning for Autonomous Underwater Vehicles Under Ocean Current Disturbance,” IEEE Trans. on Intelligent Vehicles, vol. 8, no. 1, pp. 108–120, 2023. https://doi.org/10.1109/TIV.2022.3153352.

R. S. Sutton and A. G. Barto, Reinforcement Learning: An Introduction. The MIT Press, second ed., 2018.

V. Mnih, K. Kavukcuoglu, D. Silver, A. Graves, I. Antonoglou, D. Wierstra, and M. Riedmiller, “Playing Atari with Deep Reinforcement Learning,” 2013. https://doi.org/10.48550/arXiv.1312.5602.

B. Garau, A. Alvarez, and G. Oliver, “AUV navigation through turbulent ocean environments supported by onboard H-ADCP,” in IEEE Int. Conf. on Robotics and Automation, pp. 3556–3561, 2006. https://doi.org/10.1109/ROBOT.2006.1642245.

A. Mansfield, D. G. Macharet, and M. A. Hsieh, “Energy-efficient orienteering problem in the presence of ocean currents,” in IEEE/RSJ Int. Conf. on Intelligent Robots and Systems, pp. 10081–10087, 2022. https://doi.org/10.1109/IROS47612.2022.9981818.

Published

2026-03-14

How to Cite

Alves Neto, A., Costa da Silva Campos, V., & Guimarães Macharet, D. (2026). Energy-efficient Navigation in Unknown Static Flow Fields using Reinforcement Learning. IEEE Latin America Transactions, 24(4), 328–337. Retrieved from https://latamt.ieeer9.org/index.php/transactions/article/view/10321