Abstract
Whether the products on the shelf are neatly displayed is one of the important factors affecting consumers’ desire to buy, which is more obvious in the toy industry. In order to detect the neatness of the toy commodities on the shelf, we propose a method based on texture recognition to measure the neatness of the commodities. Since the whole of neat toy commodities will show regular texture features, in the field of computer vision, how to extract effective texture features from image parameters has always been the focus and difficulty of texture analysis. In this paper, we use a multiquadratic kernel modeling learnable local histogram layer to extract effective texture features and use the convolutional block attention module to filter out features that are more conducive to classification, proposed a new network named histogram residual attention network (HRANet). In the test performance stage, we use the proposed HRANet to test on DTD, MINC-2500, GTOS-mobile, and KTH-T2b datasets, and the accuracy is significantly higher than the original HistRes. Finally, we use the features obtained by HRANet to represent the features of product display and use the cosine distance to measure the similarity of product regions, which effectively quantifies the neatness of product display.
Similar content being viewed by others
References
Feng, Q., Liu, J., Gong, J.: UAV remote sensing for urban vegetation mapping using random forest and texture analysis. Remote Sens. 7, 1074–1094 (2015)
Selvan, S., Ramakrishnan, S.: SVD-based modeling for image texture classification using wavelet transformation. IEEE Trans. Image Process. 16, 2688–2696 (2007)
Bruno, A., Collorec, R., Bézy-Wendling, J., et al.: Texture analysis in medical imaging. In: Contemporary Perspectives in Three-Dimensional Biomedical Imaging, pp. 133–164 (1997)
Andrearczyk, V., Whelan, P.F.: Deep learning in texture analysis and its application to tissue image classification. In: Depeursinge, A., Omar, S., Al, K., Mitchell, J.R. (eds.) Biomedical Texture Analysis, pp. 95–129. Academic Press, Cambridge (2017)
Peeples, J., Xu, W., Zare, A.: Histogram Layers for Texture Analysis. arXiv:2001.00215 (2020)
Liu, L., Chen, J., Fieguth, P., et al.: From BoW to CNN: two decades of texture representation for texture classification. Int. J. Comput. Vis. 127, 74–109 (2019)
Cimpoi, M., Maji, S., Vedaldi, A.: Deep filter banks for texture recognition and segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3828–3836 (2015)
Song, Y., Zhang, F., Li, Q, et al.: Locally-transferred fisher vectors for texture classification. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4912–4920 (2017)
Zhang, H., Xue, J., Dana, K.: Deep ten: texture encoding network. In: Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pp. 708–717 (2017)
Xue, J., Zhang, H., Dana, K.: Deep texture manifold for ground terrain recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 558–567 (2018)
Nguyen, VL., Vu, NS., Gosselin, PH.: A handcrafted normalized-convolution network for texture classification. In: Proceedings of the IEEE International Conference on Computer Vision Workshops, pp. 1238–1245 (2017)
Hu, Y., Long, Z., AlRegib, G.: Multi-level texture encoding and representation (multer) based on deep neural networks. In: 2019 IEEE International Conference on Image Processing (ICIP), pp. 4410–4414 (2019)
Zhai, W., Cao, Y., Zhang, J, et al.: Deep multiple-attribute-perceived network for real-world texture recognition. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 3613–3622 (2019)
Zhai, W., Cao, Y., Zha, ZJ, et al.: Deep structure-revealed network for texture recognition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11010–11019 (2020)
Tao, Z., Wei, T., Li, J.: Wavelet multi-level attention capsule network for texture classification. IEEE Signal Process. Lett. 28, 1215–1219 (2021)
Fradi, H., Fradi, A., Dugelay, J.L.: Multi-layer feature fusion and selection from convolutional neural networks for texture classification. In: VISIGRAPP (4: VISAPP), pp. 574–581 (2021)
Chen, Z., Li, F., Quan, Y, et al.: Deep texture recognition via exploiting cross-layer statistical self-similarity. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5231–5240 (2021)
Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), pp. 1:886–893 (2005)
Wang, Z., Li, H., Ouyang, W, et al.: Learnable histogram: statistical context features for deep neural networks. In: European Conference on Computer Vision, pp. 246–262 (2016)
Sedighi, V., Fridrich, J.: Histogram layer, moving convolutional neural networks towards feature-based steganalysis. Electron. Imaging 2017, 50–55 (2017)
Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7132–7141 (2018)
Woo, S., Park, J., Lee, J.Y., et al.: CBAM: convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3–19 (2018)
Cimpoi, M., Maji, S., Kokkinos, I., et al.: Describing textures in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3606–3613 (2014)
Bell, S., Upchurch, P., Snavely, N., et al.: Material recognition in the wild with the materials in context database. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3479–3487 (2015)
Caputo, B., Hayman, E., Mallikarjuna, P.: Class-specific material categorisation. In: Tenth IEEE International Conference on Computer Vision (ICCV’05), vol. 1. IEEE, pp. 2: 1597–1604 (2005)
Anzai, Y.: Pattern Recognition and Machine Learning. Elsevier, Amsterdam (2012)
Acknowledgements
This work was partly supported by National Natural Science Foundation of China (61772198, U20A20228).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Zang, Y., Ding, C., Hu, W. et al. HRANet: histogram-residual-attention network used to measure neatness of toy placement. SIViP 17, 295–303 (2023). https://doi.org/10.1007/s11760-022-02232-0
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11760-022-02232-0