{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,4,13]],"date-time":"2025-04-13T00:53:11Z","timestamp":1744505591577},"reference-count":35,"publisher":"MDPI AG","issue":"24","license":[{"start":{"date-parts":[[2020,12,18]],"date-time":"2020-12-18T00:00:00Z","timestamp":1608249600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Sensors"],"abstract":"Countries around the world have paid increasing attention to the issue of marine security, and sea target detection is a key task to ensure marine safety. Therefore, it is of great significance to propose an efficient and accurate sea-surface target detection algorithm. The anchor-setting method of the traditional YOLO v3 only uses the degree of overlap between the anchor and the ground-truth box as the standard. As a result, the information of some feature maps cannot be used, and the required accuracy of target detection is hard to achieve in a complex sea environment. Therefore, two new anchor-setting methods for the visual detection of sea targets were proposed in this paper: the average method and the select-all method. In addition, cross PANet, a feature fusion structure for cross-feature maps was developed and was used to obtain a better baseline cross YOLO v3, where different anchor-setting methods were combined with a focal loss for experimental comparison in the datasets of sea buoys and existing sea ships, SeaBuoys and SeaShips, respectively. The results showed that the method proposed in this paper could significantly improve the accuracy of YOLO v3 in detecting sea-surface targets, and the highest value of mAP in the two datasets is 98.37% and 90.58%, respectively.<\/jats:p>","DOI":"10.3390\/s20247263","type":"journal-article","created":{"date-parts":[[2020,12,21]],"date-time":"2020-12-21T06:01:08Z","timestamp":1608530468000},"page":"7263","source":"Crossref","is-referenced-by-count":29,"title":["Study on Visual Detection Algorithm of Sea Surface Targets Based on Improved YOLOv3"],"prefix":"10.3390","volume":"20","author":[{"given":"Tao","family":"Liu","sequence":"first","affiliation":[{"name":"College of Intelligent Systems Science and Engineering, Harbin Engineering University, Harbin 150001, China"}]},{"given":"Bo","family":"Pang","sequence":"additional","affiliation":[{"name":"College of Intelligent Systems Science and Engineering, Harbin Engineering University, Harbin 150001, China"}]},{"given":"Shangmao","family":"Ai","sequence":"additional","affiliation":[{"name":"College of Shipbuilding Engineering, Harbin Engineering University, Harbin 150001, China"}]},{"given":"Xiaoqiang","family":"Sun","sequence":"additional","affiliation":[{"name":"College of Intelligent Systems Science and Engineering, Harbin Engineering University, Harbin 150001, China"}]}],"member":"1968","published-online":{"date-parts":[[2020,12,18]]},"reference":[{"key":"ref_1","unstructured":"Krizhevsky, A., Sutskever, I., and Hinton, G. (2012). ImageNet Classification with Deep Convolutional Neural Networks. Adv. Neural Inf. Process. Syst., 25."},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., and Rabinovich, A. (2014, January 23\u201328). Going Deeper with Convolutions. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA.","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Anhoucke, V.V., Ioffe, S., Shlens, J., and Wojna, Z. (2015). Rethinking the inception architecture for computer vision. arXiv.","DOI":"10.1109\/CVPR.2016.308"},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Ioffe, S., Vanhoucke, V., and Alemi, A. (2016). Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. arXiv.","DOI":"10.1609\/aaai.v31i1.11231"},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (2016, January 27\u201330). Deep Residual Learning for Image Recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Xie, S., Girshick, R., Doll\u00e1r, P., Tu, Z., and He, K. (2017, January 21\u201326). Aggregated Residual Transformations for Deep Neural Networks. Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.634"},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Van Der Maaten, L., and Weinberger, K.Q. (2016). Densely Connected Convolutional Networks. arXiv.","DOI":"10.1109\/CVPR.2017.243"},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Wang, C.Y., Liao, H.Y.M., Yeh, I.H., Wu, Y.H., Chen, P.Y., and Hsieh, J.W. (2019). CSPNet: A New Backbone that can Enhance Learning Capability of CNN. arXiv.","DOI":"10.1109\/CVPRW50498.2020.00203"},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Doll\u00e1r, P., Girshick, R., He, K., Hariharan, B., and Belongie, S. (2016). Feature Pyramid Networks for Object Detection. arXiv.","DOI":"10.1109\/CVPR.2017.106"},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Liu, S., Qi, L., Qin, H., Shi, J., and Jia, J. (2018, January 18\u201323). Path Aggregation Network for Instance Segmentation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00913"},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Tan, M., Pang, R., and Le, Q.V. (2019). EfficientDet: Scalable and Efficient Object Detection. arXiv.","DOI":"10.1109\/CVPR42600.2020.01079"},{"key":"ref_12","unstructured":"Liu, S., Huang, D., and Wang, Y. (2019). Learning Spatial Fusion for Single-Shot Object Detection. arXiv."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Girshick, R., Donahue, J., Darrell, T., and Malik, J. (2014, January 23\u201328). Rich feature hierarchies for accurate object detection and semantic segmentation. Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA.","DOI":"10.1109\/CVPR.2014.81"},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Girshick, R. (2015). Fast R-CNN. arXiv.","DOI":"10.1109\/ICCV.2015.169"},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","article-title":"Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks","volume":"39","author":"Ren","year":"2017","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., and Farhadi, A. (2015). You only look once: Unified, real-time object detection. arXiv.","DOI":"10.1109\/CVPR.2016.91"},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Redmon, J., and Farhadi, A. (2017, January 21\u201326). YOLO9000: Better, Faster, Stronger. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.690"},{"key":"ref_18","unstructured":"Redmon, J., and Farhadi, A. (2018). YOLOv3: An Incremental Improvement. arXiv."},{"key":"ref_19","unstructured":"Bochkovskiy, A., Wang, C.Y., and Liao, H.Y.M. (2020). YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., and Berg, A.C. (2016). SSD: Single shot multibox detector. European Conference on Computer Vision, Springer.","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Cui, H., Yang, Y., Liu, M., Shi, T., and Qi, Q. (2019, January 17\u201320). Ship Detection: An Improved YOLOv3 Method. Proceedings of the OCEANS 2019\u2014Marseille, Marseille, France.","DOI":"10.1109\/OCEANSE.2019.8867209"},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"751","DOI":"10.1109\/LGRS.2018.2882551","article-title":"Squeeze and Excitation Rank Faster R-CNN for Ship Detection in SAR Images","volume":"16","author":"Lin","year":"2019","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"e5820","DOI":"10.1002\/cpe.5820","article-title":"Improved region convolutional neural network for ship detection in multiresolution synthetic aperture radar images","volume":"32","author":"Xiao","year":"2020","journal-title":"Concurr. Comput. Pract. Exp."},{"key":"ref_24","first-page":"1","article-title":"Comparison of two deep learning methods for ship target recognition with optical remotely sensed data","volume":"1","author":"Zhang","year":"2020","journal-title":"Neural Comput. Appl."},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Liu, Z., Hu, J., Weng, L., and Yang, Y. (2017, January 17\u201320). Rotated region based CNN for ship detection. Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China.","DOI":"10.1109\/ICIP.2017.8296411"},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"3614","DOI":"10.1080\/01431161.2019.1706781","article-title":"Learning a robust CNN-based rotation insensitive model for ship detection in VHR remote sensing images","volume":"41","author":"Dong","year":"2020","journal-title":"Int. J. Remote Sens."},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Huang, Z., Sui, B., Wen, J., and Jiang, G. (2020). An Intelligent Ship Image\/Video Detection and Classification Method with Improved Regressive Deep Convolutional Neural Network. Complexity, 2020.","DOI":"10.1155\/2020\/1520872"},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Cao, C., Wu, J., Zeng, X., Feng, Z., Wang, T., Yan, X., Wu, Z., Wu, Q., and Huang, Z. (2020). Research on Airplane and Ship Detection of Aerial Remote Sensing Images Based on Convolutional Neural Network. Sensors, 20.","DOI":"10.3390\/s20174696"},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Li, Y., Guo, J., Guo, X., Liu, K., Zhao, W., Luo, Y., and Wang, Z. (2020). A Novel Target Detection Method of the Unmanned Surface Vehicle under All-Weather Conditions with an Improved YOLOV3. Sensors, 20.","DOI":"10.3390\/s20174885"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Dai, W., Mao, Y., Yuan, R., Liu, Y., Pu, X., and Li, C. (2020). A Novel Detector Based on Convolution Neural Networks for Multiscale SAR Ship Detection in Complex Background. Sensors, 20.","DOI":"10.3390\/s20092547"},{"key":"ref_31","doi-asserted-by":"crossref","first-page":"104812","DOI":"10.1016\/j.ssci.2020.104812","article-title":"Deep learning for autonomous ship-oriented small ship detection","volume":"130","author":"Chen","year":"2020","journal-title":"Saf. Sci."},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Su, H., Wei, S., Wang, M., Zhou, L., Shi, J., and Zhang, X. (2019, January 26\u201329). Ship Detection Based on RetinaNet-Plus for High-Resolution SAR Imagery. Proceedings of the 2019 6th Asia-Pacific Conference on Synthetic Aperture Radar (APSAR), Xiamen, China.","DOI":"10.1109\/APSAR46974.2019.9048269"},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"2593","DOI":"10.1109\/TMM.2018.2865686","article-title":"SeaShips: A Large-Scale Precisely Annotated Dataset for Ship Detection","volume":"20","author":"Shao","year":"2018","journal-title":"IEEE Trans. Multimed."},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Rezatofighi, H., Tsoi, N., Gwak, J., Sadeghian, A., Reid, I., and Savarese, S. (2019, January 15\u201321). Generalized Intersection Over Union: A Metric and a Loss for Bounding Box Regression. Proceedings of the 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00075"},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"He, T., Zhang, Z., Zhang, H., Zhang, Z., Xie, J., and Li, M. (2019, January 15\u201321). Bag of Tricks for Image Classification with Convolutional Neural Networks. Proceedings of the 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00065"}],"container-title":["Sensors"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1424-8220\/20\/24\/7263\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,7,6]],"date-time":"2024-07-06T14:18:24Z","timestamp":1720275504000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1424-8220\/20\/24\/7263"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,12,18]]},"references-count":35,"journal-issue":{"issue":"24","published-online":{"date-parts":[[2020,12]]}},"alternative-id":["s20247263"],"URL":"https:\/\/doi.org\/10.3390\/s20247263","relation":{},"ISSN":["1424-8220"],"issn-type":[{"value":"1424-8220","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,12,18]]}}}