Reduced Image Classes in Modified U-Net for Mars Rover Navigation
DOI:
https://doi.org/10.47611/jsrhs.v14i1.8651Keywords:
Image Segmentation, Mars Rovers, Mars, NavigationAbstract
Rover navigation currently relies on algorithms to automatically determine their path. This is because the distance from Earth to Mars means that real time communication is impossible. Current navigation algorithms have difficulties in identifying terrain, causing problems such as becoming stuck in soft terrain. Furthermore, the available computation power and memory are limited on a rover. This paper presents both a modified U-Net model to identify parts of the terrain and combining multiple classes to have less output classes. The proposed method was to combine classes like soil and bedrock into more generalized classes, like traversable and untraversable, to reduce memory usage and needed computational power. Combining the classes shows that a model can be trained faster, and in some cases even improve. Testing this method on low resolution images has shown improved results in testing. After training, a three-class model is able to yield a higher mIoU of 0.4583 on a test set compared to the full five-class model, which achieved 0.3451. This method is non-specific to U-Net and can be applied to many different models. Combining this method with other models and larger datasets during training could be an option of improving the accuracy of models running on less processing power, allowing for use on platforms such as Mars rovers.
Downloads
References or Bibliography
Swan, M. R., Atha, D., Leopold, H. A., Gildner, M., Oij, S., Chiu, C., & Ono, M. (2021). AI4MARS: A dataset for terrain-aware autonomous driving on Mars. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). https://doi.org/10.1109/cvprw53098.2021.00226
Verma, V., Maimone, M. W., Gaines, D. M., Francis, R., Estlin, T. A., Kuhn, S. R., Rabideau, G. R., Chien, S. A., McHenry, M. M., Graser, E. J., Rankin, A. L., & Thiel, E. R. (2023). Autonomous robotics is driving Perseverance rover's progress on Mars. Science robotics, 8(80), eadi3099. https://doi.org/10.1126/scirobotics.adi3099
Rankin, A., Maimone, M., Biesiadecki, J., Patel, N., Levine, D., & Toupet, O. (2020). Driving curiosity: Mars rover mobility trends during the first seven years. 2020 IEEE Aerospace Conference. https://doi.org/10.1109/aero47225.2020.9172469
Lee, D.-H., & Liu, J.-L. (2023). Multi-task UNet architecture for end-to-end autonomous driving. arXiv [Cs.LG]. Retrieved from http://arxiv.org/abs/2112.08967
Geiger, A., Lenz, P., & Urtasun, R. (2012). Are we ready for autonomous driving? The Kitti Vision Benchmark Suite. 2012 IEEE Conference on Computer Vision and Pattern Recognition, 3354–3361. https://doi.org/10.1109/cvpr.2012.6248074
Ronneberger, O., Fischer, P., & Brox, T. (2015). U-Net: Convolutional Networks for Biomedical Image Segmentation. Lecture Notes in Computer Science, 234–241. https://doi.org/10.1007/978-3-319-24574-4_28
Feng, L., Wang, S., Wang, D., Xiong, P., Xie, J., Hu, Y., Zhang, M., Wu, E. Q., & Song, A. (2023). Mobile-deeprfb: A lightweight terrain classifier for automatic Mars Rover Navigation. IEEE Transactions on Automation Science and Engineering, 1–10. https://doi.org/10.1109/tase.2023.3340190
Liu, S., Huang, D., & Wang, Y. (2018). Receptive field block net for accurate and fast object detection. Lecture Notes in Computer Science, 404–419. https://doi.org/10.1007/978-3-030-01252-6_24
Kay, S., Quoos, M., Field, R., Prokopczyk, M. J., De Benedetti, M., Mohammad, F., ... & Ntagiou, E. V. (2023). AI-enabled Computer Vision Framework for Automated Knowledge Extraction in Planetary Rover Operations. Submitted to ASTRA.
Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., & Salakhutdinov, R. (2014). Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research, 15(1), 1929-1958.
Published
How to Cite
Issue
Section
Copyright (c) 2025 Roy Qiu; Victoria Lloyd

This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
Copyright holder(s) granted JSR a perpetual, non-exclusive license to distriute & display this article.


