{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,31]],"date-time":"2025-03-31T04:51:02Z","timestamp":1743396662953,"version":"3.37.3"},"reference-count":31,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["41930301","41761080"],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Pattern Recognition Letters"],"published-print":{"date-parts":[[2021,1]]},"DOI":"10.1016\/j.patrec.2020.11.014","type":"journal-article","created":{"date-parts":[[2020,12,10]],"date-time":"2020-12-10T22:25:17Z","timestamp":1607639117000},"page":"45-53","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":20,"special_numbering":"C","title":["Multi-focus image fusion algorithm based on supervised learning for fully convolutional neural network"],"prefix":"10.1016","volume":"141","author":[{"given":"Heng","family":"Li","sequence":"first","affiliation":[]},{"given":"Liming","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Meirong","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"Yulong","family":"Li","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.patrec.2020.11.014_bib0001","doi-asserted-by":"crossref","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","article-title":"Multi-focus image fusion with a deep convolutional neural network","volume":"36","author":"Liu","year":"2017","journal-title":"Inf. Fusion"},{"issue":"5","key":"10.1016\/j.patrec.2020.11.014_bib0002","doi-asserted-by":"crossref","first-page":"347","DOI":"10.1049\/iet-ipr.2014.0311","article-title":"Simultaneous image fusion and denoising with adaptive sparse representation","volume":"9","author":"Liu","year":"2014","journal-title":"IET Image Process"},{"issue":"2","key":"10.1016\/j.patrec.2020.11.014_bib0003","doi-asserted-by":"crossref","first-page":"16","DOI":"10.1007\/s41651-019-0039-9","article-title":"Scene classification of high-resolution remotely sensed image based on resnet","volume":"3","author":"Wang","year":"2019","journal-title":"J. Geovisualiz. Spat. Anal."},{"issue":"3","key":"10.1016\/j.patrec.2020.11.014_bib0004","article-title":"Multifocus image fusion using phase congruency","volume":"24","author":"Zhan","year":"2015","journal-title":"J. Electron. Image."},{"issue":"4","key":"10.1016\/j.patrec.2020.11.014_bib0005","doi-asserted-by":"crossref","first-page":"532","DOI":"10.1109\/TCOM.1983.1095851","article-title":"The Laplacian pyramid as a compact image code","volume":"31","author":"Burt","year":"1983","journal-title":"IEEE Trans. Commun."},{"issue":"3","key":"10.1016\/j.patrec.2020.11.014_bib0006","doi-asserted-by":"crossref","first-page":"480","DOI":"10.1364\/JOSAA.35.000480","article-title":"Multi-focus image fusion algorithm based on Laplacian pyramids","volume":"35","author":"Sun","year":"2018","journal-title":"J. Opt. Soc. Am. A"},{"issue":"3","key":"10.1016\/j.patrec.2020.11.014_bib0007","doi-asserted-by":"crossref","first-page":"235","DOI":"10.1006\/gmip.1995.1022","article-title":"Multisensor image fusion using the wavelet transform","volume":"57","author":"Li","year":"1995","journal-title":"Graph. Models Image Process."},{"key":"10.1016\/j.patrec.2020.11.014_bib0008","doi-asserted-by":"crossref","first-page":"9","DOI":"10.1016\/j.sigpro.2013.10.010","article-title":"Region level based multi-focus image fusion using quaternion wavelet and normalized cut","volume":"97","author":"Liu","year":"2014","journal-title":"Signal Process"},{"issue":"7","key":"10.1016\/j.patrec.2020.11.014_bib0009","doi-asserted-by":"crossref","first-page":"1334","DOI":"10.1016\/j.sigpro.2009.01.012","article-title":"Multifocus image fusion using the nonsubsampled contourlet transform","volume":"89","author":"Zhang","year":"2009","journal-title":"Signal Process"},{"key":"10.1016\/j.patrec.2020.11.014_bib0010","doi-asserted-by":"crossref","first-page":"68","DOI":"10.1016\/j.infrared.2018.05.006","article-title":"General fusion method for infrared and visual images via latent low-rank representation and local non-subsampled shearlet transform","volume":"92","author":"Cheng","year":"2018","journal-title":"Infrared Phys. Technol."},{"key":"10.1016\/j.patrec.2020.11.014_bib0011","doi-asserted-by":"crossref","DOI":"10.1007\/s11760-018-1303-z","article-title":"Multimodal medical image fusion using non-subsampled shearlet transform and pulse coupled neural network incorporated with morphological gradient","author":"Ramlal","year":"2018","journal-title":"Signal Image Video Process"},{"issue":"7","key":"10.1016\/j.patrec.2020.11.014_bib0012","doi-asserted-by":"crossref","first-page":"2864","DOI":"10.1109\/TIP.2013.2244222","article-title":"Image fusion with guided filtering","volume":"22","author":"Li","year":"2013","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"10.1016\/j.patrec.2020.11.014_bib0013","doi-asserted-by":"crossref","first-page":"147","DOI":"10.1016\/j.inffus.2011.07.001","article-title":"Image matting for fusion of multi-focus images in dynamic scenes","volume":"14","author":"Li","year":"2013","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.patrec.2020.11.014_bib0014","doi-asserted-by":"crossref","first-page":"139","DOI":"10.1016\/j.inffus.2014.05.004","article-title":"Multi-focus image fusion with dense SIFT","volume":"23","author":"Liu","year":"2015","journal-title":"Inf. Fusion"},{"issue":"2","key":"10.1016\/j.patrec.2020.11.014_bib0015","doi-asserted-by":"crossref","first-page":"91","DOI":"10.1016\/j.optcom.2011.08.078","article-title":"Multifocus image fusion and denoising scheme based on homogeneity similarity","volume":"285","author":"Li","year":"2012","journal-title":"Opt. Commun."},{"key":"10.1016\/j.patrec.2020.11.014_bib0016","doi-asserted-by":"crossref","first-page":"26","DOI":"10.1016\/j.imavis.2019.03.001","article-title":"Multi-scale convolutional neural network for multi-focus image fusion","volume":"85","author":"Mustafa","year":"2019","journal-title":"Image Vis. Comput."},{"key":"10.1016\/j.patrec.2020.11.014_bib0017","doi-asserted-by":"crossref","first-page":"3845","DOI":"10.1109\/TIP.2020.2966075","article-title":"Unsupervised deep image fusion with structure tensor representations","volume":"29","author":"Jung","year":"2020","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.patrec.2020.11.014_bib0018","doi-asserted-by":"crossref","first-page":"60","DOI":"10.1016\/j.inffus.2013.11.005","article-title":"Multi-scale weighted gradient-based fusion for multi-focus images","volume":"20","author":"Zhou","year":"2014","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.patrec.2020.11.014_bib0019","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"3431","article-title":"Fully convolutional networks for semantic segmentation","author":"Long","year":"2015"},{"key":"10.1016\/j.patrec.2020.11.014_bib0020","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"770","article-title":"Deep residual learning for image recognition","author":"He","year":"2016"},{"key":"10.1016\/j.patrec.2020.11.014_bib0021","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"4700","article-title":"Densely connected convolutional networks","author":"Huang","year":"2017"},{"journal-title":"Network in network","year":"2013","author":"Lin","key":"10.1016\/j.patrec.2020.11.014_bib0022"},{"key":"10.1016\/j.patrec.2020.11.014_bib0023","unstructured":"\u201cThe PASCAL Visual Object Classes Challenge 2007 (VOC2007).\u201d http:\/\/host.robots.ox.ac.uk\/pascal\/VOC\/voc2007\/ (accessed Jul. 11, 2019)."},{"issue":"7","key":"10.1016\/j.patrec.2020.11.014_bib0024","doi-asserted-by":"crossref","first-page":"1334","DOI":"10.1016\/j.sigpro.2009.01.012","article-title":"Multifocus image fusion using the nonsubsampled contourlet transform","volume":"89","author":"Zhang","year":"2009","journal-title":"Signal Process"},{"issue":"2","key":"10.1016\/j.patrec.2020.11.014_bib0025","doi-asserted-by":"crossref","first-page":"119","DOI":"10.1016\/j.inffus.2005.09.006","article-title":"Pixel-and region-based image fusion with complex wavelets","volume":"8","author":"Lewis","year":"2007","journal-title":"Inf. Fusion"},{"issue":"1","key":"10.1016\/j.patrec.2020.11.014_bib0026","doi-asserted-by":"crossref","first-page":"49","DOI":"10.1109\/TIM.2018.2838778","article-title":"Medical image fusion with parameter-adaptive pulse coupled neural network in nonsubsampled shearlet transform domain","volume":"68","author":"Yin","year":"2018","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"2","key":"10.1016\/j.patrec.2020.11.014_bib0027","doi-asserted-by":"crossref","first-page":"74","DOI":"10.1016\/j.inffus.2010.03.002","article-title":"Performance comparison of different multi-resolution transforms for image fusion","volume":"12","author":"Li","year":"2011","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.patrec.2020.11.014_bib0028","unstructured":"\u201cMansour Nejati | Lytro Multi-focus Dataset.\u201d https:\/\/mansournejati.ece.iut.ac.ir\/content\/lytro-multi-focus-dataset (accessed Jul. 11, 2019)."},{"issue":"4","key":"10.1016\/j.patrec.2020.11.014_bib0029","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1109\/TIP.2003.819861","article-title":"others, \u201cImage quality assessment: from error visibility to structural similarity","volume":"13","author":"Wang","year":"2004","journal-title":"IEEE Trans. Image Process."},{"issue":"9","key":"10.1016\/j.patrec.2020.11.014_bib0030","doi-asserted-by":"crossref","first-page":"2137","DOI":"10.1109\/TIP.2009.2025006","article-title":"A total variation-based algorithm for pixel-level image fusion","volume":"18","author":"Kumar","year":"2009","journal-title":"IEEE Trans. Image Process."},{"issue":"3","key":"10.1016\/j.patrec.2020.11.014_bib0031","doi-asserted-by":"crossref","first-page":"81","DOI":"10.1109\/97.995823","article-title":"A universal image quality index","volume":"9","author":"Wang","year":"2002","journal-title":"IEEE Signal Process. Lett."}],"container-title":["Pattern Recognition Letters"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0167865520304256?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0167865520304256?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2021,1,9]],"date-time":"2021-01-09T17:29:35Z","timestamp":1610213375000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0167865520304256"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,1]]},"references-count":31,"alternative-id":["S0167865520304256"],"URL":"https:\/\/doi.org\/10.1016\/j.patrec.2020.11.014","relation":{},"ISSN":["0167-8655"],"issn-type":[{"type":"print","value":"0167-8655"}],"subject":[],"published":{"date-parts":[[2021,1]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Multi-focus image fusion algorithm based on supervised learning for fully convolutional neural network","name":"articletitle","label":"Article Title"},{"value":"Pattern Recognition Letters","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.patrec.2020.11.014","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2020 Elsevier B.V. All rights reserved.","name":"copyright","label":"Copyright"}]}}