QS-Hyper: A Quality-Sensitive Hyper Network for the No-Reference Image Quality Assessment | SpringerLink
Skip to main content

QS-Hyper: A Quality-Sensitive Hyper Network for the No-Reference Image Quality Assessment

  • Conference paper
  • First Online:
Neural Information Processing (ICONIP 2021)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 13111))

Included in the following conference series:

  • 2119 Accesses

Abstract

Blind/no-reference image quality assessment (IQA) aims to provide a quality score for a single image without references. In this context, deep learning models can capture various image artifacts, which made significant progress in this study. However, current IQA methods generally utilize the pre-trained convolution neural networks (CNNs) on classification tasks to obtain image representations, which do not perfectly represent the quality of images. In order to solve this problem, this paper uses semi-supervised representation learning to train a quality-sensitive encoder (QS-encoder), which can extract image features specifically for image quality. Intuitively, this feature is more conducive to train the IQA model than the feature used for classification tasks. Thus, QS-encoder is plunged into a carefully designed hyper network to build a quality-sensitive hyper network (QS-hyper) to solve IQA tasks in more general and complex environments. Extensive experiments on the public IQA datasets show that our method outperformed most state-of-art methods on both Pearson linear correlation coefficient (PLCC) and Spearman’s rank correlation coefficient (SRCC), and it made 3% PLCC improvement and 3.9% SRCC improvement on TID2013 datasets. Therefore, it proves that our method is superior in capturing various image distortions, which meets a broader range of evaluation requirements.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
JPY 3498
Price includes VAT (Japan)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
JPY 11439
Price includes VAT (Japan)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
JPY 14299
Price includes VAT (Japan)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Bosse, S., Maniry, D., Müller, K.R., Wiegand, T., Samek, W.: Deep neural networks for no-reference and full-reference image quality assessment. IEEE Trans. Image Process. 27(1), 206–219 (2017)

    Google Scholar 

  2. Bromley, J., Guyon, I., LeCun, Y., Säckinger, E., Shah, R.: Signature verification using a “siamese" time delay neural network. Adv. Neural Inf. Process. Syst. 6, 737–744 (1993)

    Google Scholar 

  3. Caron, M., Bojanowski, P., Joulin, A., Douze, M.: Deep clustering for unsupervised learning of visual features. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 132–149 (2018)

    Google Scholar 

  4. Chen, X., He, K.: Exploring simple siamese representation learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 15750–15758 (2021)

    Google Scholar 

  5. Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: a large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition, pp. 248–255. IEEE (2009)

    Google Scholar 

  6. Ghadiyaram, D., Bovik, A.C.: Massive online crowdsourced study of subjective and objective picture quality. IEEE Trans. Image Process. 25(1), 372–387 (2015)

    Google Scholar 

  7. He, K., Fan, H., Wu, Y., Xie, S., Girshick, R.: Momentum contrast for unsupervised visual representation learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9729–9738 (2020)

    Google Scholar 

  8. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778 (2016)

    Google Scholar 

  9. Hjelm, R.D., et al.: Learning deep representations by mutual information estimation and maximization. arXiv preprint arXiv:1808.06670 (2018)

  10. Hosu, V., Lin, H., Sziranyi, T., Saupe, D.: Koniq-10k: an ecologically valid database for deep learning of blind image quality assessment. IEEE Trans. Image Process. 29, 4041–4056 (2020)

    Google Scholar 

  11. Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 25, 1097–1105 (2012)

    Google Scholar 

  12. Lin, K.Y., Wang, G.: Hallucinated-iqa: no-reference image quality assessment via adversarial learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 732–741 (2018)

    Google Scholar 

  13. Liu, X., Van De Weijer, J., Bagdanov, A.D.: Rankiqa: learning from rankings for no-reference image quality assessment. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1040–1049 (2017)

    Google Scholar 

  14. Mittal, A., Moorthy, A.K., Bovik, A.C.: No-reference image quality assessment in the spatial domain. IEEE Transactions on Image Processing 21(12), 4695–4708 (2012)

    Google Scholar 

  15. Moorthy, A.K., Bovik, A.C.: A two-step framework for constructing blind image quality indices. IEEE Signal Process. Lett. 17(5), 513–516 (2010)

    Google Scholar 

  16. Ou, F.Z., et al.: Sdd-fiqa: unsupervised face image quality assessment with similarity distribution distance. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7670–7679 (2021)

    Google Scholar 

  17. Ponomarenko, N., et al.: Image database tid2013: peculiarities, results and perspectives. Signal Process. Image Commun. 30, 57–77 (2015)

    Google Scholar 

  18. Su, S., et al.: Blindly assess image quality in the wild guided by a self-adaptive hyper network. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3667–3676 (2020)

    Google Scholar 

  19. Xue, W., Mou, X., Zhang, L., Bovik, A.C., Feng, X.: Blind image quality assessment using joint statistics of gradient magnitude and laplacian features. IEEE Trans. Image Process. 23(11), 4850–4862 (2014)

    Google Scholar 

  20. Zhang, L., Zhang, L., Bovik, A.C.: A feature-enriched completely blind image quality evaluator. IEEE Trans. Image Process. 24(8), 2579–2591 (2015)

    Google Scholar 

  21. Zhang, X., Yu, W., Jiang, N., Zhang, Y., Zhang, Z.: Sps: A subjective perception score for text-to-image synthesis. In: 2021 IEEE International Symposium on Circuits and Systems (ISCAS), pp. 1–5. IEEE (2021)

    Google Scholar 

  22. Zhang, X., Zhang, Y., Zhang, Z., Yu, W., Jiang, N., He, G.: Deep feature compatibility for generated images quality assessment. In: Yang, H., Pasupa, K., Leung, A.CS., Kwok, J.T., Chan, J.H., King, I. (eds.) Neural Information Processing. ICONIP 2020. Communications in Computer and Information Science, vol. 1332. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-63820-7_40

  23. Zhang, Y., Zhang, X., Zhang, Z., Yu, W., Jiang, N., He, G.: No-reference quality assessment based on spatial statistic for generated images. In: Yang, H., Pasupa, K., Leung, A.CS., Kwok, J.T., Chan, J.H., King, I. (eds.) Neural Information Processing. ICONIP 2020. Communications in Computer and Information Science, vol. 1332. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-63820-7_57

  24. Zhu, H., Li, L., Wu, J., Dong, W., Shi, G.: Metaiqa: deep meta-learning for no-reference image quality assessment. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 14143–14152 (2020)

    Google Scholar 

Download references

Acknowledgement

This research is supported by Sichuan Science and Technology Program (No. 2020YFS0307, No. 2020YFG0430, No. 2019YFS0146), Mianyang Science and Technology Program (2020YFZJ016).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Wenxin Yu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zhang, X., Zhang, Y., Yu, W., Nie, L., Jiang, N., Gong, J. (2021). QS-Hyper: A Quality-Sensitive Hyper Network for the No-Reference Image Quality Assessment. In: Mantoro, T., Lee, M., Ayu, M.A., Wong, K.W., Hidayanto, A.N. (eds) Neural Information Processing. ICONIP 2021. Lecture Notes in Computer Science(), vol 13111. Springer, Cham. https://doi.org/10.1007/978-3-030-92273-3_26

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-92273-3_26

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-92272-6

  • Online ISBN: 978-3-030-92273-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics