HRANet: histogram-residual-attention network used to measure neatness of toy placement | Signal, Image and Video Processing Skip to main content
Log in

HRANet: histogram-residual-attention network used to measure neatness of toy placement

  • Original Paper
  • Published:
Signal, Image and Video Processing Aims and scope Submit manuscript

Abstract

Whether the products on the shelf are neatly displayed is one of the important factors affecting consumers’ desire to buy, which is more obvious in the toy industry. In order to detect the neatness of the toy commodities on the shelf, we propose a method based on texture recognition to measure the neatness of the commodities. Since the whole of neat toy commodities will show regular texture features, in the field of computer vision, how to extract effective texture features from image parameters has always been the focus and difficulty of texture analysis. In this paper, we use a multiquadratic kernel modeling learnable local histogram layer to extract effective texture features and use the convolutional block attention module to filter out features that are more conducive to classification, proposed a new network named histogram residual attention network (HRANet). In the test performance stage, we use the proposed HRANet to test on DTD, MINC-2500, GTOS-mobile, and KTH-T2b datasets, and the accuracy is significantly higher than the original HistRes. Finally, we use the features obtained by HRANet to represent the features of product display and use the cosine distance to measure the similarity of product regions, which effectively quantifies the neatness of product display.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price includes VAT (Japan)

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9

Similar content being viewed by others

References

  1. Feng, Q., Liu, J., Gong, J.: UAV remote sensing for urban vegetation mapping using random forest and texture analysis. Remote Sens. 7, 1074–1094 (2015)

    Article  Google Scholar 

  2. Selvan, S., Ramakrishnan, S.: SVD-based modeling for image texture classification using wavelet transformation. IEEE Trans. Image Process. 16, 2688–2696 (2007)

    Article  MathSciNet  Google Scholar 

  3. Bruno, A., Collorec, R., Bézy-Wendling, J., et al.: Texture analysis in medical imaging. In: Contemporary Perspectives in Three-Dimensional Biomedical Imaging, pp. 133–164 (1997)

  4. Andrearczyk, V., Whelan, P.F.: Deep learning in texture analysis and its application to tissue image classification. In: Depeursinge, A., Omar, S., Al, K., Mitchell, J.R. (eds.) Biomedical Texture Analysis, pp. 95–129. Academic Press, Cambridge (2017)

    Chapter  Google Scholar 

  5. Peeples, J., Xu, W., Zare, A.: Histogram Layers for Texture Analysis. arXiv:2001.00215 (2020)

  6. Liu, L., Chen, J., Fieguth, P., et al.: From BoW to CNN: two decades of texture representation for texture classification. Int. J. Comput. Vis. 127, 74–109 (2019)

    Article  Google Scholar 

  7. Cimpoi, M., Maji, S., Vedaldi, A.: Deep filter banks for texture recognition and segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3828–3836 (2015)

  8. Song, Y., Zhang, F., Li, Q, et al.: Locally-transferred fisher vectors for texture classification. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4912–4920 (2017)

  9. Zhang, H., Xue, J., Dana, K.: Deep ten: texture encoding network. In: Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pp. 708–717 (2017)

  10. Xue, J., Zhang, H., Dana, K.: Deep texture manifold for ground terrain recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 558–567 (2018)

  11. Nguyen, VL., Vu, NS., Gosselin, PH.: A handcrafted normalized-convolution network for texture classification. In: Proceedings of the IEEE International Conference on Computer Vision Workshops, pp. 1238–1245 (2017)

  12. Hu, Y., Long, Z., AlRegib, G.: Multi-level texture encoding and representation (multer) based on deep neural networks. In: 2019 IEEE International Conference on Image Processing (ICIP), pp. 4410–4414 (2019)

  13. Zhai, W., Cao, Y., Zhang, J, et al.: Deep multiple-attribute-perceived network for real-world texture recognition. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 3613–3622 (2019)

  14. Zhai, W., Cao, Y., Zha, ZJ, et al.: Deep structure-revealed network for texture recognition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11010–11019 (2020)

  15. Tao, Z., Wei, T., Li, J.: Wavelet multi-level attention capsule network for texture classification. IEEE Signal Process. Lett. 28, 1215–1219 (2021)

    Article  Google Scholar 

  16. Fradi, H., Fradi, A., Dugelay, J.L.: Multi-layer feature fusion and selection from convolutional neural networks for texture classification. In: VISIGRAPP (4: VISAPP), pp. 574–581 (2021)

  17. Chen, Z., Li, F., Quan, Y, et al.: Deep texture recognition via exploiting cross-layer statistical self-similarity. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5231–5240 (2021)

  18. Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), pp. 1:886–893 (2005)

  19. Wang, Z., Li, H., Ouyang, W, et al.: Learnable histogram: statistical context features for deep neural networks. In: European Conference on Computer Vision, pp. 246–262 (2016)

  20. Sedighi, V., Fridrich, J.: Histogram layer, moving convolutional neural networks towards feature-based steganalysis. Electron. Imaging 2017, 50–55 (2017)

    Article  Google Scholar 

  21. Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7132–7141 (2018)

  22. Woo, S., Park, J., Lee, J.Y., et al.: CBAM: convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3–19 (2018)

  23. Cimpoi, M., Maji, S., Kokkinos, I., et al.: Describing textures in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3606–3613 (2014)

  24. Bell, S., Upchurch, P., Snavely, N., et al.: Material recognition in the wild with the materials in context database. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3479–3487 (2015)

  25. Caputo, B., Hayman, E., Mallikarjuna, P.: Class-specific material categorisation. In: Tenth IEEE International Conference on Computer Vision (ICCV’05), vol. 1. IEEE, pp. 2: 1597–1604 (2005)

  26. Anzai, Y.: Pattern Recognition and Machine Learning. Elsevier, Amsterdam (2012)

    MATH  Google Scholar 

Download references

Acknowledgements

This work was partly supported by National Natural Science Foundation of China (61772198, U20A20228).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Wenjun Hu.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zang, Y., Ding, C., Hu, W. et al. HRANet: histogram-residual-attention network used to measure neatness of toy placement. SIViP 17, 295–303 (2023). https://doi.org/10.1007/s11760-022-02232-0

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11760-022-02232-0

Keywords

Navigation