Two guidance joint network based on coarse map and edge map for camouflaged object detection | Applied Intelligence Skip to main content
Log in

Two guidance joint network based on coarse map and edge map for camouflaged object detection

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Camouflaged object detection (COD) entails identifying objects in an image that blend with the background. However, most traditional COD methods have not comprehensively considered the information provided by the overall region and edges of the objects. To address this problem, a new two-guidance joint network based on coarse and edge maps is proposed for COD. Particularly, an information guidance module is designed to inject edges and overall information into the network’s backbone features. Meanwhile, a feature observation model based on skip connections and multi-scale perception is designed to capture multi-scale image details and structures. To avoid the loss of semantic information in low-level features, a full-image attention mechanism is designed to integrate high-level features into low-level features, thereby improving the resolution of the object masks. We compared the proposed network with state-of-the-art models on three well-known datasets, and the experimental results show the proposed network has significant improvement. By exploring valuable boundary information and overall object information, the proposed network can segment object edges while also considering the segmentation effect of the entire object. Our code has been open-sourced at https://github.com/Huah2019/TJNet.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price includes VAT (Japan)

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Data Availability

The training datasets can be found in the https://drive.google.com/file/d/1Kifp7I0n9dlWKXXNIbN7kgyokoRY4Yz7/view?usp=sharing. The test datasets can be found in the https://drive.google.com/file/d/1SLRB5Wg1Hdy7CQ74s3mTQ3ChhjFRSFdZ/view?usp=sharing. All the datasets is sourced from Github’s open-source project and can be accessed through https://github.com/thograce/BGNet.

References

  1. Cuthill IC, Stevens M, Sheppard J, Maddocks T, Párraga CA, Troscianko TS (2005) Disruptive coloration and background pattern matching. Nature 434(7029):72–74

    Article  Google Scholar 

  2. Price N, Green S, Troscianko J, Tregenza T, Stevens M (2019) Background matching and disruptive coloration as habitat-specific strategies for camouflage. Sci Rep 9(1):7840

    Article  Google Scholar 

  3. Sun Y, Wang S, Chen C, Xiang T-Z (2022) Boundary-guided camouflaged object detection. Preprint arXiv:2207.00794

  4. Fan D-P, Ji G-P, Sun G, Cheng M-M, Shen J, Shao L (2020) Camouflaged object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2777–2787

  5. Lu Yao ZB, Wei W et al (2020) Design and experiment of belt screen seed sorter for camellia oleifera fused with photoelectric color sorting technology. Trans Chin Soc Agric Mach 51(S1):429–439

    Google Scholar 

  6. Peijun C (2014) Research on recognizing tea-leaves and impurities based on image processing and pattern classification. Nanjing University of Aeronautics and Astronautics, In Nanjing, pp 1–2

    Google Scholar 

  7. Singh V, Misra AK (2017) Detection of plant leaf diseases using image segmentation and soft computing techniques. Inf Process Agric 4(1):41–49

    Google Scholar 

  8. Pérez-de la Fuente R, Delclòs X, Peñalver E, Speranza M, Wierzchos J, Ascaso C, Engel MS (2012) Early evolution and ecology of camouflage in insects. Proc Nat Acad Sci 109(52):21414–21419

    Article  Google Scholar 

  9. Fan D-P, Ji G-P, Zhou T, Chen G, Fu H, Shen J, Shao L (2020) Pranet: Parallel reverse attention network for polyp segmentation. In International conference on medical image computing and computer-assisted intervention, pp 263–273. Springer

  10. Ge S, Jin X, Ye Q, Luo Z, Li Q (2018) Image editing by object-aware optimal boundary searching and mixed-domain composition. Comput Vis Media 4:71–82

    Article  Google Scholar 

  11. Chu H-K, Hsu W-H, Mitra NJ, Cohen-Or D, Wong T-T, Lee T-Y (2010) Camouflage images. ACM Trans Graph 29(4):51–1

    Article  Google Scholar 

  12. Mei H, Ji G-P, Wei Z, Yang X, Wei X, Fan D-P (2021) Camouflaged object segmentation with distraction mining. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 8772–8781

  13. Yan J, Le T-N, Nguyen K-D, Tran M-T, Do T-T, Nguyen TV (2021) Mirrornet: Bio-inspired camouflaged object segmentation. IEEE Access 9:43290–43300

    Article  Google Scholar 

  14. Du X, Xu X, Ma K (2022) Icgnet: Integration context-based reverse-contour guidance network for polyp segmentation. In: Proceedings of the international joint conferences on artificial intelligence, pp 877–883

  15. Xu X, Zhu M, Yu J, Chen S, Hu X, Yang Y (2021) Boundary guidance network for camouflage object detection. Image Vis Comput 114:104283

    Article  Google Scholar 

  16. Liang Y, Qin G, Sun M, Wang X, Yan J, Zhang Z (2023) A systematic review of image-level camouflaged object detection with deep learning. Neurocomputing, p 127050

  17. Ren J, Hu X, Zhu L, Xu X, Xu Y, Wang W, Deng Z, Heng P-A (2021) Deep texture-aware features for camouflaged object detection. IEEE Trans Circuits Syst Vid Technol

  18. Fan D-P, Ji G-P, Cheng M-M, Shao L (2021) Concealed object detection. IEEE Trans Pattern Anal Mach Intell 44(10):6024–6042

    Article  Google Scholar 

  19. Zhou T, Zhou Y, Gong C, Yang J, Zhang Y (2022) Feature aggregation and propagation network for camouflaged object detection. IEEE Trans Image Process 31:7036–7047

    Article  Google Scholar 

  20. Zhai Q, Li X, Yang F, Chen C, Cheng H, Fan D-P (2021) Mutual graph learning for camouflaged object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 12997–13007

  21. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. Advances in neural information processing systems, 30

  22. Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S et al (2020) An image is worth 16x16 words: Transformers for image recognition at scale. Preprint arXiv:2010.11929

  23. Pei J, Cheng T, Fan D-P, Tang H, Chen C, Van Gool L (2022) Osformer: One-stage camouflaged instance segmentation with transformers. In: European conference on computer vision, pp 19–37. Springer

  24. Xing H, Wang Y, Wei X, Tang H, Gao S, Zhang W (2023) Go closer to see better: Camouflaged object detection via object area amplification and figure-ground conversion. IEEE Trans Circuits Syst Video Technol

  25. Zhong Y, Li B, Tang L, Kuang S, Wu S, Ding S (2022) Detecting camouflaged object in frequency domain. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 4504–4513

  26. He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778

  27. Yang F, Zhai Q, Li X, Huang R, Luo A, Cheng H, Fan D-P (2021) Uncertainty-guided transformer reasoning for camouflaged object detection. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 4146–4155

  28. Lin T-Y, Dollár P, Girshick R, He K, Hariharan B, Belongie S (2017) Feature pyramid networks for object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2117–2125

  29. Zhang Q, Ge Y, Zhang C, Bi H (2023) Tprnet: camouflaged object detection via transformer-induced progressive refinement network. Vis Comput 39(10):4593–4607

    Article  Google Scholar 

  30. Zhao J-X, Liu J-J, Fan D-P, Cao Y, Yang J, Cheng M-M (2019) Egnet: Edge guidance network for salient object detection. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 8779–8788

  31. Xiao J, Chen T, Hu X, Zhang G, Wang S (2023) Boundary-guided context-aware network for camouflaged object detection. Neural Comput Appl 35(20):15075–15093

    Article  Google Scholar 

  32. Gao S-H, Cheng M-M, Zhao K, Zhang X-Y, Yang M-H, Torr P (2019) Res2net: A new multi-scale backbone architecture. IEEE Trans Pattern Anal Mach Intell 43(2):652–662

    Article  Google Scholar 

  33. Wang Q, Wu B, Zhu P, Li P, Zuo W, Hu Q (2020) Eca-net: Efficient channel attention for deep convolutional neural networks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 11534–11542

  34. Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7132–7141

  35. Xie E, Wang W, Wang W, Ding M, Shen C, Luo P (2020) Segmenting transparent objects in the wild. In: Computer vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XIII 16, pp 696–711. Springer

  36. Le T-N, Nguyen TV, Nie Z, Tran M-T, Sugimoto A (2019) Anabranch network for camouflaged object segmentation. Comput Vis Image Understand 184:45–56

    Article  Google Scholar 

  37. Skurowski P, Abdulameer H,Błaszczyk J, Depta T, Kornacki A, Kozieł P (2018) Animal camouflage analysis: Chameleon database. Unpublished manuscript 2(6):7

  38. Perazzi F, Krähenbühl P, Pritch Y, Hornung A (2012) Saliency filters: Contrast based filtering for salient region detection. In: 2012 IEEE conference on computer vision and pattern recognition, pp 733–740. IEEE

  39. Margolin R, Zelnik-Manor L, Tal A (2014) How to evaluate foreground maps? In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 248–255

  40. Fan D-P, Ji G-P, Qin X, Cheng M-M (2021) Cognitive vision inspired object segmentation metric and loss function. Sci Sin Inf 6(6):5

    Google Scholar 

  41. Fan D-P, Cheng M-M, Liu Y, Li T, Borji A (2017) Structure-measure: A new way to evaluate foreground maps. In: Proceedings of the IEEE international conference on computer vision, pp 4548–4557

Download references

Author information

Authors and Affiliations

Authors

Contributions

All authors have equal contributions.

Corresponding author

Correspondence to Fang Qi.

Ethics declarations

Competing interests

The authors declare that they have no competing interests.

Ethics approval

We declare that this manuscript is original, has not been published before and is not currently being considered for publication elsewhere.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

This work is supported by the National Key Research and Development Program (No. 2022YFD2101101), the Project of Scientifc and Technological Innovation Planning of Hunan Province (No.2021NK1020), the earmarked fund for China Agriculture Research System (CARS-19), and the High Performance Computing Center of Central South University.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Tang, Z., Tang, J., Zou, D. et al. Two guidance joint network based on coarse map and edge map for camouflaged object detection. Appl Intell 54, 7531–7544 (2024). https://doi.org/10.1007/s10489-024-05559-y

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-024-05559-y

Keywords

Navigation