{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2023,12,28]],"date-time":"2023-12-28T06:57:23Z","timestamp":1703746643832},"reference-count":63,"publisher":"Springer Science and Business Media LLC","license":[{"start":{"date-parts":[[2022,6,21]],"date-time":"2022-06-21T00:00:00Z","timestamp":1655769600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,6,21]],"date-time":"2022-06-21T00:00:00Z","timestamp":1655769600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61802111, 61872125"],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100011447","name":"Science and Technology Department of Henan Province","doi-asserted-by":"publisher","award":["201300210400, 212102210094"],"id":[{"id":"10.13039\/501100011447","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"DOI":"10.1007\/s10489-022-03847-z","type":"journal-article","created":{"date-parts":[[2022,6,21]],"date-time":"2022-06-21T08:03:32Z","timestamp":1655798612000},"update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["LDN-RC: a lightweight denoising network with residual connection to improve adversarial robustness"],"prefix":"10.1007","author":[{"given":"Xiuli","family":"Chai","sequence":"first","affiliation":[]},{"given":"Tongtong","family":"Wei","sequence":"additional","affiliation":[]},{"given":"Zhen","family":"Chen","sequence":"additional","affiliation":[]},{"ORCID":"http:\/\/orcid.org\/0000-0002-2867-589X","authenticated-orcid":false,"given":"Xin","family":"He","sequence":"additional","affiliation":[]},{"given":"Zhihua","family":"Gan","sequence":"additional","affiliation":[]},{"given":"Xiangjun","family":"Wu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,6,21]]},"reference":[{"key":"3847_CR1","doi-asserted-by":"publisher","unstructured":"Lee S, Lee H, Yoon S (2020) Adversarial vertex mixup: toward better adversarially robust generalization. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 269\u2013278. https:\/\/doi.org\/10.1109\/CVPR42600.2020.00035","DOI":"10.1109\/CVPR42600.2020.00035"},{"key":"3847_CR2","doi-asserted-by":"publisher","unstructured":"Deng Z, Yang X, Xu S et al (2021) LiBRe: a practical bayesian approach to adversarial detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 972\u2013982. https:\/\/doi.org\/10.1109\/cvpr46437.2021.00103","DOI":"10.1109\/cvpr46437.2021.00103"},{"key":"3847_CR3","doi-asserted-by":"publisher","unstructured":"Qiu S, Liu Q, Zhou S, Wu C (2019) Review of artificial intelligence adversarial attack and defense technologies. Appl Sci 9(5). https:\/\/doi.org\/10.3390\/app9050909","DOI":"10.3390\/app9050909"},{"key":"3847_CR4","unstructured":"Madry A, Makelov A, Schmidt L et al (2018) Towards deep learning models resistant to adversarial attacks. In: International conference on learning representations, pp 1\u201328"},{"key":"3847_CR5","doi-asserted-by":"crossref","unstructured":"Liao F, Liang M, Dong Y et al (2018) Defense against adversarial attacks using high-level representation guided denoiser. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1778\u20131787","DOI":"10.1109\/CVPR.2018.00191"},{"key":"3847_CR6","doi-asserted-by":"publisher","unstructured":"Ma C, Ying L (2021) Achieving adversarial robustness requires an active teacher. J Comput Math 39(6):880\u2013896. https:\/\/doi.org\/10.4208\/jcm.2105-m2020-0310","DOI":"10.4208\/jcm.2105-m2020-0310"},{"key":"3847_CR7","doi-asserted-by":"publisher","DOI":"10.1007\/s10489-021-02759-8","author":"S Wang","year":"2021","unstructured":"Wang S, Gong Y (2021) Adversarial example detection based on saliency map features. Appl Intell. https:\/\/doi.org\/10.1007\/s10489-021-02759-8","journal-title":"Appl Intell"},{"key":"3847_CR8","doi-asserted-by":"crossref","unstructured":"Xie C, Wu Y, Maaten L, Van Der et al (2019) Feature denoising for improving adversarial robustness. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 501\u2013509","DOI":"10.1109\/CVPR.2019.00059"},{"issue":"7","key":"3847_CR9","doi-asserted-by":"publisher","first-page":"3187","DOI":"10.1109\/TII.2018.2822680","volume":"14","author":"Z Cui","year":"2018","unstructured":"Cui Z, Xue F, Cai X et al (2018) Detection of malicious code variants based on deep learning. IEEE Trans Industr Inf 14(7):3187\u20133196. https:\/\/doi.org\/10.1109\/TII.2018.2822680","journal-title":"IEEE Trans Industr Inf"},{"key":"3847_CR10","doi-asserted-by":"crossref","unstructured":"Mustafa A, Khan S, Hayat M et al (2019) Adversarial defense by restricting the hidden space of deep neural networks. In: Proceedings of the international conference on computer vision, pp 3384\u20133393","DOI":"10.1109\/ICCV.2019.00348"},{"key":"3847_CR11","doi-asserted-by":"publisher","unstructured":"Wadlow LR (2017) MagNet: a two-pronged defense against adversarial examples. In: Proceedings of the 24th ACM-SIGSAC conference on computer and communications security (ACM CCS), pp 135\u2013147. https:\/\/doi.org\/10.1145\/3133956.3134057","DOI":"10.1145\/3133956.3134057"},{"issue":"5","key":"3847_CR12","doi-asserted-by":"publisher","first-page":"635","DOI":"10.1109\/JPROC.2021.3050042","volume":"109","author":"G Ortiz-Jimenez","year":"2021","unstructured":"Ortiz-Jimenez G, Modas A, Moosavi-Dezfooli SM, Frossard P (2021) Optimism in the face of adversity: understanding and improving deep learning through adversarial robustness. Proc IEEE 109(5):635\u2013659. https:\/\/doi.org\/10.1109\/JPROC.2021.3050042","journal-title":"Proc IEEE"},{"key":"3847_CR13","doi-asserted-by":"publisher","first-page":"568","DOI":"10.1016\/j.ins.2020.08.043","volume":"547","author":"T Li","year":"2021","unstructured":"Li T, Liu A, Liu X et al (2021) Understanding adversarial robustness via critical attacking route. Inf Sci 547:568\u2013578. https:\/\/doi.org\/10.1016\/j.ins.2020.08.043","journal-title":"Inf Sci"},{"key":"3847_CR14","doi-asserted-by":"publisher","DOI":"10.1007\/s10489-021-02371-w","author":"X Fang","year":"2021","unstructured":"Fang X, Li Z, Yang G (2021) A novel approach to generating high-resolution adversarial examples. Appl Intell. https:\/\/doi.org\/10.1007\/s10489-021-02371-w","journal-title":"Appl Intell"},{"key":"3847_CR15","doi-asserted-by":"crossref","unstructured":"Naseer M, Khan S, Hayat M et al (2020) A self-supervised approach for adversarial robustness. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 259\u2013268","DOI":"10.1109\/CVPR42600.2020.00034"},{"key":"3847_CR16","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2021.107141","author":"L Wang","year":"2021","unstructured":"Wang L, Chen X, Tang R et al (2021) Improving adversarial robustness of deep neural networks by using semantic information. Knowl Based Syst. https:\/\/doi.org\/10.1016\/j.knosys.2021.107141","journal-title":"Knowl Based Syst"},{"key":"3847_CR17","doi-asserted-by":"crossref","unstructured":"Ghosh P, Losalka A, Black MJ (2019) Resisting adversarial attacks using Gaussian mixture variational autoencoders. In: Proceedings of the AAAI conference on artificial intelligence, pp 541\u2013548","DOI":"10.1609\/aaai.v33i01.3301541"},{"issue":"10","key":"3847_CR18","doi-asserted-by":"publisher","first-page":"1","DOI":"10.3390\/e23101359","volume":"23","author":"K Mahmood","year":"2021","unstructured":"Mahmood K, Gurevin D, van Dijk M, Ha Nguyen P (2021) Beware the black-box: On the robustness of recent defenses to adversarial examples. Entropy 23(10):1\u201340. https:\/\/doi.org\/10.3390\/e23101359","journal-title":"Entropy"},{"issue":"10","key":"3847_CR19","doi-asserted-by":"publisher","first-page":"1453","DOI":"10.1002\/int.22258","volume":"35","author":"Z Yin","year":"2020","unstructured":"Yin Z, Wang H, Wang J et al (2020) Defense against adversarial attacks by low-level image transformations. Int J Intell Syst 35(10):1453\u20131466. https:\/\/doi.org\/10.1002\/int.22258","journal-title":"Int J Intell Syst"},{"key":"3847_CR20","doi-asserted-by":"publisher","unstructured":"Liu N, Du M, Guo R et al (2020) Adversarial attacks and defenses: an interpretation perspective. https:\/\/doi.org\/10.1145\/3468507.3468519","DOI":"10.1145\/3468507.3468519"},{"key":"3847_CR21","doi-asserted-by":"publisher","unstructured":"Nesti F, Biondi A, Buttazzo G (2021) Detecting adversarial examples by input transformations, defense perturbations, and voting. IEEE Trans Neural Netw Learn Syst 11\u201313. https:\/\/doi.org\/10.1109\/tnnls.2021.3105238","DOI":"10.1109\/tnnls.2021.3105238"},{"key":"3847_CR22","doi-asserted-by":"crossref","unstructured":"Vivek BS, Babu RV (2020)Single-step adversarial training with dropout scheduling. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 947\u2013956","DOI":"10.1109\/CVPR42600.2020.00103"},{"issue":"4","key":"3847_CR23","doi-asserted-by":"publisher","first-page":"1513","DOI":"10.1109\/TDSC.2020.3024660","volume":"18","author":"W Wei","year":"2021","unstructured":"Wei W, Liu L (2021) Robust deep learning ensemble against deception. IEEE Trans Dependable Secur Comput 18(4):1513\u20131527. https:\/\/doi.org\/10.1109\/TDSC.2020.3024660","journal-title":"IEEE Trans Dependable Secur Comput"},{"key":"3847_CR24","doi-asserted-by":"crossref","unstructured":"He Z, Rakin AS, Fan D (2019) Parametric noise injection: Trainable randomness to improve deep neural network robustness against adversarial attack. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 588\u2013597","DOI":"10.1109\/CVPR.2019.00068"},{"key":"3847_CR25","doi-asserted-by":"crossref","unstructured":"Jeddi A, Shafiee MJ, Karg M et al (2020) Learn2Perturb: An end-to-end feature perturbation learning to improve adversarial robustness. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1238\u20131247","DOI":"10.1109\/CVPR42600.2020.00132"},{"key":"3847_CR26","doi-asserted-by":"publisher","first-page":"195","DOI":"10.1016\/j.neucom.2018.04.027","volume":"307","author":"U Shaham","year":"2018","unstructured":"Shaham U, Yamada Y, Negahban S (2018) Understanding adversarial training: Increasing local stability of supervised models through robust optimization. Neurocomputing 307:195\u2013204. https:\/\/doi.org\/10.1016\/j.neucom.2018.04.027","journal-title":"Neurocomputing"},{"key":"3847_CR27","doi-asserted-by":"crossref","unstructured":"Chen T, Liu S, Chang S et al (2020) Adversarial robustness: from self-supervised pre-training to fine-tuning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 696\u2013705","DOI":"10.1109\/CVPR42600.2020.00078"},{"key":"3847_CR28","doi-asserted-by":"crossref","unstructured":"Chen P (2017) ZOO: zeroth order optimization based black-box attacks to deep neural networks without training substitute models. In: Proceedings of the 10th ACM workshop on artificial intelligence and security, pp 15\u201326","DOI":"10.1145\/3128572.3140448"},{"key":"3847_CR29","doi-asserted-by":"crossref","unstructured":"Wu T, Liu Z, Huang Q et al (2021) Adversarial robustness under long-tailed distribution. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8659\u20138668","DOI":"10.1109\/CVPR46437.2021.00855"},{"key":"3847_CR30","doi-asserted-by":"publisher","DOI":"10.1007\/s10489-021-02523-y","author":"J Ho","year":"2021","unstructured":"Ho J, Lee BG, Kang DK (2021)Attack-less adversarial training for a robust adversarial defense. Appl Intell. https:\/\/doi.org\/10.1007\/s10489-021-02523-y","journal-title":"Appl Intell"},{"key":"3847_CR31","doi-asserted-by":"crossref","unstructured":"Awasthi P, Yu G, Ferng C-S et al (2020) Adversarial robustness across representation spaces. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7604\u20137612","DOI":"10.1109\/CVPR46437.2021.00752"},{"key":"3847_CR32","doi-asserted-by":"crossref","unstructured":"Li G, Ding S, Luo J, Liu C (2020) Enhancing intrinsic adversarial robustness via feature pyramid decoder. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 797\u2013805","DOI":"10.1109\/CVPR42600.2020.00088"},{"key":"3847_CR33","doi-asserted-by":"crossref","unstructured":"Cheng M, Chen P-Y, Liu S et al (2021)Self-progressing robust training. In: Proceedings of the AAAI conference on artificial intelligence, pp 7107\u20137115","DOI":"10.1609\/aaai.v35i8.16874"},{"key":"3847_CR34","doi-asserted-by":"crossref","unstructured":"Cazenavette G, Murdock C, Lucey S (2021) Architectural adversarial robustness: the case for deep pursuit. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7150\u20137158","DOI":"10.1109\/CVPR46437.2021.00707"},{"key":"3847_CR35","unstructured":"Zhang H, Yu Y, Jiao J et al (2019) Theoretically principled trade-off between robustness and accuracy. In: Proceedings of the 36th international conference on machine learning (ICML), pp 12907\u201312929"},{"key":"3847_CR36","doi-asserted-by":"crossref","unstructured":"Dong Y, Liao F, Pang T et al (2018) Boosting adversarial attacks with momentum. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 9185\u20139193","DOI":"10.1109\/CVPR.2018.00957"},{"key":"3847_CR37","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2020.107309","author":"Y Shi","year":"2020","unstructured":"Shi Y, Han Y, Zhang Q, Kuang X (2020) Adaptive iterative attack towards explainable adversarial robustness. Pattern Recogn. https:\/\/doi.org\/10.1016\/j.patcog.2020.107309","journal-title":"Pattern Recogn"},{"key":"3847_CR38","doi-asserted-by":"crossref","unstructured":"Carlini N (2017) Towards evaluating the robustness of neural networks. In: Proceedings of the 38th IEEE symposium on security and privacy (SP), pp 39\u201357","DOI":"10.1109\/SP.2017.49"},{"key":"3847_CR39","unstructured":"Fawzi A, Frossard P (2016) DeepFool: a simple and accurate method to fool deep neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2574\u20132582"},{"key":"3847_CR40","unstructured":"Goodfellow IJ (2017) Adversarial examples in the physical world. In: Proceedings of the 5th international conference on learning representations (ICLR), pp 1\u201314"},{"key":"3847_CR41","doi-asserted-by":"crossref","unstructured":"Xie C, Zhang Z, Wang J et al (2019) Improving transferability of adversarial examples with input diversity. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2725\u20132734","DOI":"10.1109\/CVPR.2019.00284"},{"key":"3847_CR42","doi-asserted-by":"publisher","first-page":"515","DOI":"10.1109\/TSP.2020.3045206","volume":"69","author":"Y Jin","year":"2021","unstructured":"Jin Y, Lai L (2021) On the adversarial robustness of hypothesis testing. IEEE Trans Signal Process 69:515\u2013530. https:\/\/doi.org\/10.1109\/TSP.2020.3045206","journal-title":"IEEE Trans Signal Process"},{"key":"3847_CR43","doi-asserted-by":"crossref","unstructured":"Huang B, Ke Z, Wang Y et al (2021) Adversarial defence by diversified simultaneous training of deep ensembles. In: Proceedings of the AAAI conference on artificial intelligence, pp 7823\u20137831","DOI":"10.1609\/aaai.v35i9.16955"},{"key":"3847_CR44","unstructured":"Li X, Li X, Pan D, Zhu D (2020) Improving adversarial robustness via probabilistically compact loss with logit constraints. In: Proceedings of the AAAI conference on artificial intelligence, pp 8482\u20138490"},{"key":"3847_CR45","doi-asserted-by":"crossref","unstructured":"Addepalli S, Vivek BS, Baburaj A et al (2020) Towards achieving adversarial robustness by enforcing feature consistency across bit planes. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1017\u20131026","DOI":"10.1109\/CVPR42600.2020.00110"},{"key":"3847_CR46","doi-asserted-by":"publisher","unstructured":"Hlihor P, Volpi R, Malag\u00f2 L (2020) Evaluating the robustness of defense mechanisms based on autoencoder reconstructions against carlini-wagner adversarial attacks. In: Proceedings of the northern lights deep learning workshop. https:\/\/doi.org\/10.7557\/18.5173","DOI":"10.7557\/18.5173"},{"key":"3847_CR47","unstructured":"Deng Z, Zhang L, Ghorbani A, Zou J (2020) Improving adversarial robustness via unlabeled out-of-domain data. In: Proceedings of the 24th international conference on artificial intelligence and statistics (AISTATS)"},{"key":"3847_CR48","doi-asserted-by":"publisher","first-page":"1291","DOI":"10.1109\/TIP.2020.3042083","volume":"30","author":"C Zhang","year":"2021","unstructured":"Zhang C, Liu A, Liu X et al (2021) Interpreting and improving adversarial robustness of deep neural networks with neuron sensitivity. IEEE Trans Image Process 30:1291\u20131304. https:\/\/doi.org\/10.1109\/TIP.2020.3042083","journal-title":"IEEE Trans Image Process"},{"key":"3847_CR49","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.neunet.2021.02.023","volume":"140","author":"M Tavakoli","year":"2021","unstructured":"Tavakoli M, Agostinelli F, Baldi P (2021) SPLASH: learnable activation functions for improving accuracy and adversarial robustness. Neural Netw 140:1\u201312. https:\/\/doi.org\/10.1016\/j.neunet.2021.02.023","journal-title":"Neural Netw"},{"key":"3847_CR50","doi-asserted-by":"publisher","first-page":"5769","DOI":"10.1109\/TIP.2021.3082317","volume":"30","author":"A Liu","year":"2021","unstructured":"Liu A, Liu X, Yu H et al (2021) Training robust deep neural networks via adversarial noise propagation. IEEE Trans Image Process 30:5769\u20135781. https:\/\/doi.org\/10.1109\/TIP.2021.3082317","journal-title":"IEEE Trans Image Process"},{"issue":"2","key":"3847_CR51","doi-asserted-by":"publisher","first-page":"291","DOI":"10.1007\/s00521-015-1874-3","volume":"27","author":"GG Wang","year":"2016","unstructured":"Wang GG, Lu M, Dong YQ, Zhao XJ (2016)Self-adaptive extreme learning machine. Neural Comput Appl 27(2):291\u2013303. https:\/\/doi.org\/10.1007\/s00521-015-1874-3","journal-title":"Neural Comput Appl"},{"issue":"1","key":"3847_CR52","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1177\/1687814015624832","volume":"8","author":"JH Yi","year":"2016","unstructured":"Yi JH, Wang J, Wang GG (2016) Improved probabilistic neural networks with self-adaptive strategies for transformer fault diagnosis problem. Adv Mech Eng 8(1):1\u201313. https:\/\/doi.org\/10.1177\/1687814015624832","journal-title":"Adv Mech Eng"},{"key":"3847_CR53","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2021.108303","author":"K Han","year":"2022","unstructured":"Han K, Xia B, Li Y (2022) (AD)2: adversarial domain adaptation to defense with adversarial perturbation removal. Pattern Recogn. https:\/\/doi.org\/10.1016\/j.patcog.2021.108303","journal-title":"Pattern Recogn"},{"key":"3847_CR54","first-page":"1","volume":"32","author":"Z Yue","year":"2019","unstructured":"Yue Z, Yong H, Zhao Q et al (2019) Variational denoising network: toward blind noise modeling and removal. Adv Neural Inf Process Syst 32:1\u201312","journal-title":"Adv Neural Inf Process Syst"},{"issue":"7","key":"3847_CR55","doi-asserted-by":"publisher","first-page":"3142","DOI":"10.1109\/TIP.2017.2662206","volume":"26","author":"K Zhang","year":"2017","unstructured":"Zhang K, Zuo W, Chen Y et al (2017) Beyond a gaussian denoiser: residual learning of deep cnn for image denoising. IEEE Trans Image Process 26(7):3142\u20133155. https:\/\/doi.org\/10.1109\/TIP.2017.2662206","journal-title":"IEEE Trans Image Process"},{"key":"3847_CR56","doi-asserted-by":"publisher","unstructured":"Lecun Y, Bottou L, Bengio Y, Ha P (1998)Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278\u20132324. https:\/\/doi.org\/10.1109\/5.726791","DOI":"10.1109\/5.726791"},{"key":"3847_CR57","unstructured":"McCrary MB (1992) Urban multicultural trauma patients. Asha 34(4)"},{"key":"3847_CR58","unstructured":"Netzer Y, Wang T, Coates A et al (2011) Reading digits in natural images with unsupervised feature learning. In: NIPS workshop on deep learning and unsupervised feature learning"},{"key":"3847_CR59","unstructured":"Rice L, Wong E, Kolter JZ (2020) Overfitting in adversarially robust deep learning. In: Proceedings of the 37th international conference on machine learning, pp 8093\u20138104"},{"key":"3847_CR60","doi-asserted-by":"crossref","unstructured":"Goldblum M, Fowl L, Feizi S, Goldstein T (2020) Adversarially robust distillation. In: Proceedings of the AAAI conference on artificial intelligence, pp 3996\u20134003","DOI":"10.1609\/aaai.v34i04.5816"},{"issue":"7","key":"3847_CR61","doi-asserted-by":"publisher","first-page":"1995","DOI":"10.1007\/s00521-015-1923-y","volume":"31","author":"GG Wang","year":"2019","unstructured":"Wang GG, Deb S, Cui Z (2019) Monarch butterfly optimization. Neural Comput Appl 31(7):1995\u20132014. https:\/\/doi.org\/10.1007\/s00521-015-1923-y","journal-title":"Neural Comput Appl"},{"key":"3847_CR62","doi-asserted-by":"publisher","unstructured":"Yang Y, Chen H, Heidari AA, Gandomi AH (2021) Hunger games search: visions, conception, implementation, deep analysis, perspectives, and towards performance shifts. Expert Syst Appl 177(114864). https:\/\/doi.org\/10.1016\/j.eswa.2021.114864","DOI":"10.1016\/j.eswa.2021.114864"},{"key":"3847_CR63","doi-asserted-by":"publisher","unstructured":"Ahmadianfar I, Heidari AA, Gandomi AH et al (2021) RUN beyond the metaphor: an efficient optimization algorithm based on runge kutta method. Expert Syst Appl 181(115079). https:\/\/doi.org\/10.1016\/j.eswa.2021.115079","DOI":"10.1016\/j.eswa.2021.115079"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-022-03847-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-022-03847-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-022-03847-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,2,8]],"date-time":"2023-02-08T22:19:41Z","timestamp":1675894781000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-022-03847-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,6,21]]},"references-count":63,"alternative-id":["3847"],"URL":"https:\/\/doi.org\/10.1007\/s10489-022-03847-z","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"value":"0924-669X","type":"print"},{"value":"1573-7497","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,6,21]]},"assertion":[{"value":"26 May 2022","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 June 2022","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}