REALTIME NEURAL NETWORK ALGORITHM FOR FULL-FRAME MARINE SURFACE OBJECTS RECOGNITION

  • V.A. Tupikov SPE "Research and Production Enterprise "Air and Marine Electronics"
  • V.A. Pavlova SPE "Research and Production Enterprise "Air and Marine Electronics"
  • V.A. Bondarenko SPE "Research and Production Enterprise "Air and Marine Electronics"
  • N.G. Holod SPE "Research and Production Enterprise "Air and Marine Electronics"
Keywords: Unmanned marine navigation, automatic detection and recognition, artificial neural net-work, image database

Abstract

The article explores modern neural network architectures for the automatic detection and recognition of marine surface objects and obstacles of given classes throughout the full image area, applicable for execution in real or near real time on an optoelectronic vision system to au-tomate and improve the safety of civil marine navigation. A formal statement of the problem of automatic detection of objects on images is given. The state-of-the-art algorithms for detecting objects in images based on use of artificial convolutional neural networks were reviewed, their comparison was made and a reasonable choice was made in favor of the most efficient neuralnetwork architecture in terms of computational complexity to recognition accuracy. The subject area is studied, as well as publicly available databases of surface objects suitable for use in the training of algorithms using artificial neural networks. The article concluded that there is insuffi-cient labeled data for training neural network algorithms, as a result of which the authors inde-pendently collected research images and video sequences, prepared and labeled the collected data containing surface marine objects and other obstacles that represent a navigation hazard for ships. Based on the selected neural network architecture, a new neural network algorithm for automatic full-frame detection and recognition of surface objects was developed, and an artificial neural network was trained using the prepared database of images of typical objects. The resulting algorithm was tested by the authors on a validation data set, the quality of its work was estimated using various metrics, and the algorithm’s performance was measured. Conclusions are made about the necessity to expand the collected database of images of typical marine objects, further steps are proposed to improve the accuracy of the developed software and algorithmic complex and its implementation to be used in a marine optoelectronic machine vision system for automa-tion and improving the safety of civil navigation.

References

1. Goodfellow I., Bengio Y., Courville A. Deep Learning. MIT Press, 2016, 781 p.
2. Deng L. and Deep D.Yu. Learning: Methods and Applications, Foundations and Trends in Signal Processing, 2013, Vol. 7, No. 3–4, pp. 197-387.
3. Girshick R. Fast r-cnn, Proceedings of the IEEE international conference on computer vision, 2015, ppP. 1440-1448.
4. Ren S. et al. Faster r-cnn: Towards real-time object detection with region proposal networks, Advances in neural information processing systems, 2015, pp. 91-99.
5. Redmon J. et al. You only look once: Unified, real-time object detection, Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 779-788.
6. Liu W. et al. Ssd: Single shot multibox detector, European conference on computer vision, Springer, Cham, 2016, pp. 21-37.
7. Ioffe S., Szegedy C. Batch normalization: Accelerating deep network training by reducing in-ternal covariate shift, arXiv preprint arXiv:1502.03167, 2015.
8. Szegedy C. et al. Scalable, high-quality object detection, ArXiv abs/1412.1441, 2015.
9. Zhou X., Wang D., Krähenbühl P. Objects as points, ArXiv abs/1904.07850, 2019.
10. Shao Z. et al. Seaships: A large-scale precisely annotated dataset for ship detection, IEEE Transactions on Multimedia, 2018, Vol. 20, No. 10, pp. 2593-2604.
11. Prasad D.K. et al. Video processing from electro-optical sensors for object detection and tracking in a maritime environment: a survey, IEEE Transactions on Intelligent Transportation Systems, 2017, Vol. 18, No. 8, pp. 1993-2016.
12. Crowston K. Amazon Mechanical Turk: A Research Tool for Organizations and Information Systems Scholars. In: Bhattacherjee A., Fitzgerald B. (eds) Shaping the Future of ICT Re-search. Methods and Approaches. IFIP Advances in Information and Communication Tech-nology. Vol. 389. Springer, Berlin, Heidelberg, 2012.
13. Yandex Toloka url: https://toloka.yandex.ru (accessed 18 February 2020).
14. Paszke A., Gross S., Massa F. and others. PyTorch: An Imperative Style, High Performance Deep Learning Library, Advances in Neural Information Processing Systems 32, Curran Asso-ciates, Inc., 2019, pp. 8024-8035.
15. Lin T.Y. et al. Focal loss for dense object detection, Proceedings of the IEEE international conference on computer vision, 2017, pp. 2980-2988.
16. He K. et al. Deep residual learning for image recognition, Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770-778.
17. Sandler M. et al. Mobilenetv2: Inverted residuals and linear bottlenecks, Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, ppP. 4510-4520.
18. Lin T., Maire M., Belongie S.J., Bourdev L.D., Girshick R.B., Hays J., Perona P., Ramanan D., Dollár P. and Zitnick C.L. Microsoft COCO: Common Objects in Context, 2014. DOI: 10.1007/978-3-319-10602-1_48.
19. Deng J., Dong W., Socher R., Li L., Kai Li and Li Fei-Fei. ImageNet: A large-scale hierar-chical image database, 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, 2009, pp. 248-255.
20. Kingma D.P., Ba J. Adam: A method for stochastic optimization, arXiv preprint arXiv:1412.6980, 2014.
Published
2020-07-10
Section
SECTION IV. COMMUNICATION, NAVIGATION, AND GUIDANCE