Abstract
Annotating medical images for disease detection is often tedious and expensive. Moreover, the available training samples for a given task are generally scarce and imbalanced. These conditions are not conducive for learning effective deep neural models. Hence, it is common to ‘transfer’ neural networks trained on natural images to the medical image domain. However, this paradigm lacks in performance due to the large domain gap between the natural and medical image data. To address that, we propose a novel concept of Pre-text Representation Transfer (PRT). In contrast to the conventional transfer learning, which fine-tunes a source model after replacing its classification layers, PRT retains the original classification layers and updates the representation layers through an unsupervised pre-text task. The task is performed with (original, not synthetic) medical images, without utilizing any annotations. This enables representation transfer with a large amount of training data. This high-fidelity representation transfer allows us to use the resulting model as a more effective feature extractor. Moreover, we can also subsequently perform the traditional transfer learning with this model. We devise a collaborative representation based classification layer for the case when we leverage the model as a feature extractor. We fuse the output of this layer with the predictions of a model induced with the traditional transfer learning performed over our pre-text transferred model. The utility of our technique for limited and imbalanced data classification problem is demonstrated with an extensive five-fold evaluation for three large-scale models, tested for five different class-imbalance ratios for CT based COVID-19 detection. Our results show a consistent gain over the conventional transfer learning with the proposed method.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Akhtar, N., Shafait, F., Mian, A.: Efficient classification with sparsity augmented collaborative representation. Pattern Recogn. 65, 136–145 (2017)
Albelwi, S.: Survey on self-supervised learning: auxiliary pretext tasks and contrastive learning methods in imaging. Entropy 24(4), 551 (2022)
Altaf, F., Islam, S., Janjua, N.K.: A novel augmented deep transfer learning for classification of COVID-19 and other thoracic diseases from X-rays. Neural Comput. Appl. 33(20), 14037–14048 (2021)
Altaf, F., Islam, S.M., Akhtar, N.: Resetting the baseline: CT-based COVID-19 diagnosis with deep transfer learning is not as accurate as widely thought. In: 2021 Digital Image Computing: Techniques and Applications (DICTA), pp. 1–8. IEEE (2021)
Altaf, F., Islam, S.M., Akhtar, N., Janjua, N.K.: Going deep in medical image analysis: concepts, methods, challenges, and future directions. IEEE Access 7, 99540–99572 (2019)
Altaf, F., Islam, S.M., Janjua, N.K., Akhtar, N.: Boosting deep transfer learning for COVID-19 classification. In: 2021 IEEE International Conference on Image Processing (ICIP), pp. 210–214. IEEE (2021)
Ericsson, L., Gouk, H., Loy, C.C., Hospedales, T.M.: Self-supervised representation learning: introduction, advances and challenges. arXiv preprint arXiv:2110.09327 (2021)
Esteva, A., et al.: Deep learning-enabled medical computer vision. NPJ Digit. Med. 4(1), 1–9 (2021)
Fang, Y., et al.: Sensitivity of chest CT for COVID-19: comparison to RT-PCR. Radiology 296(2), E115–E117 (2020)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016)
Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700–4708 (2017)
LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436–444 (2015)
Long, C., et al.: Diagnosis of the coronavirus disease (COVID-19): RRT-PCR or CT? Eur. J. Radiol. 126, 108961 (2020)
Misra, I., Maaten, L.V.D.: Self-supervised learning of pretext-invariant representations. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6707–6717 (2020)
Pham, T.D.: A comprehensive study on classification of COVID-19 on computed tomography with pretrained convolutional neural networks. Sci. Rep. 10(1), 1–8 (2020)
Reeves, A.P., et al.: A public image database to support research in computer aided diagnosis. In: 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, pp. 3715–3718. IEEE (2009)
Roberts, M., et al.: Common pitfalls and recommendations for using machine learning to detect and prognosticate for COVID-19 using chest radiographs and CT scans. Nat. Mach. Intell. 3(3), 199–217 (2021)
Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)
Tošić, I., Frossard, P.: Dictionary learning. IEEE Signal Process. Mag. 28(2), 27–38 (2011)
Wynants, L., et al.: Prediction models for diagnosis and prognosis of COVID-19: systematic review and critical appraisal. BMJ 369 (2020)
Yan, K., Wang, X., Lu, L., Summers, R.M.: Deeplesion: automated mining of large-scale lesion annotations and universal lesion detection with deep learning. J. Med. Imaging 5(3), 036501 (2018)
Yu, X., Wang, J., Hong, Q.Q., Teku, R., Wang, S.H., Zhang, Y.D.: Transfer learning for medical images analyses: a survey. Neurocomputing 489, 230–254 (2022)
Zhang, Y., Jiang, H., Miura, Y., Manning, C.D., Langlotz, C.P.: Contrastive learning of medical visual representations from paired images and text. arXiv preprint arXiv:2010.00747 (2020)
Zhao, J., Zhang, Y., He, X., Xie, P.: COVID-CT-dataset: a CT scan dataset about COVID-19. arXiv preprint arXiv:2003.13865 (2020)
Acknowledgment
This work was supported by Australian Government Research Training Program Scholarship. Dr. Akhtar is a recipient of the Office of National Intelligence National Intelligence Postdoctoral Grant # NIPG-2021-001 funded by the Australian Government.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Altaf, F., Islam, S.M.S., Janjua, N.K., Akhtar, N. (2023). Pre-text Representation Transfer for Deep Learning with Limited and Imbalanced Data: Application to CT-Based COVID-19 Detection. In: Yan, W.Q., Nguyen, M., Stommel, M. (eds) Image and Vision Computing. IVCNZ 2022. Lecture Notes in Computer Science, vol 13836. Springer, Cham. https://doi.org/10.1007/978-3-031-25825-1_9
Download citation
DOI: https://doi.org/10.1007/978-3-031-25825-1_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-25824-4
Online ISBN: 978-3-031-25825-1
eBook Packages: Computer ScienceComputer Science (R0)