{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,12,8]],"date-time":"2024-12-08T05:08:14Z","timestamp":1733634494506,"version":"3.30.1"},"reference-count":38,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2025,4,1]],"date-time":"2025-04-01T00:00:00Z","timestamp":1743465600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2025,4,1]],"date-time":"2025-04-01T00:00:00Z","timestamp":1743465600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2025,4,1]],"date-time":"2025-04-01T00:00:00Z","timestamp":1743465600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2025,4,1]],"date-time":"2025-04-01T00:00:00Z","timestamp":1743465600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2025,4,1]],"date-time":"2025-04-01T00:00:00Z","timestamp":1743465600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2025,4,1]],"date-time":"2025-04-01T00:00:00Z","timestamp":1743465600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,4,1]],"date-time":"2025-04-01T00:00:00Z","timestamp":1743465600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100003453","name":"Natural Science Foundation of Guangdong Province","doi-asserted-by":"publisher","award":["2023A1515012717"],"id":[{"id":"10.13039\/501100003453","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2023YFF0905603","2023YFF0905602"],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62176066"],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Pattern Recognition"],"published-print":{"date-parts":[[2025,4]]},"DOI":"10.1016\/j.patcog.2024.111225","type":"journal-article","created":{"date-parts":[[2024,11,26]],"date-time":"2024-11-26T22:40:13Z","timestamp":1732660813000},"page":"111225","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["Semantic decomposition and enhancement hashing for deep cross-modal retrieval"],"prefix":"10.1016","volume":"160","author":[{"given":"Lunke","family":"Fei","sequence":"first","affiliation":[]},{"given":"Zhihao","family":"He","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5214-7114","authenticated-orcid":false,"given":"Wai Keung","family":"Wong","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5497-9478","authenticated-orcid":false,"given":"Qi","family":"Zhu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7641-2237","authenticated-orcid":false,"given":"Shuping","family":"Zhao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9554-2379","authenticated-orcid":false,"given":"Jie","family":"Wen","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.patcog.2024.111225_bib0001","doi-asserted-by":"crossref","DOI":"10.1016\/j.cosrev.2020.100336","article-title":"Comparative analysis on cross-modal information retrieval: a review","volume":"39","author":"Kaur","year":"2021","journal-title":"Comput. Sci. Rev."},{"issue":"11","key":"10.1016\/j.patcog.2024.111225_bib0002","doi-asserted-by":"crossref","first-page":"5427","DOI":"10.1109\/TIP.2016.2607421","article-title":"Large-scale cross-modality search via collective matrix factorization hashing","volume":"25","author":"Ding","year":"2016","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.patcog.2024.111225_bib0003","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2020.107479","article-title":"Joint and individual matrix factorization hashing for large-scale cross-modal retrieval","volume":"107","author":"Wang","year":"2020","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.patcog.2024.111225_bib0004","doi-asserted-by":"crossref","DOI":"10.1109\/TSMC.2024.3373612","article-title":"Efficient discriminative hashing for cross-modal retrieval","author":"Huang","year":"2024","journal-title":"IEEE Trans. Syst. Man Cybern. Syst."},{"key":"10.1016\/j.patcog.2024.111225_bib0005","doi-asserted-by":"crossref","first-page":"5343","DOI":"10.1109\/TIP.2022.3195059","article-title":"Joint specifics and consistency hash learning for large-scale cross-modal retrieval","volume":"31","author":"Qin","year":"2022","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.patcog.2024.111225_bib0006","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2022.108823","article-title":"Asymmetric cross\u2013modal hashing with high\u2013level semantic similarity","volume":"130","author":"Yang","year":"2022","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.patcog.2024.111225_bib0007","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"3232","article-title":"Deep cross-modal hashing","author":"Jiang","year":"2017"},{"key":"10.1016\/j.patcog.2024.111225_bib0008","series-title":"Proceedings of the International Conference on Multimedia Retrieval","first-page":"631","article-title":"MSSPQ: multiple semantic structure-preserving quantization for cross-modal retrieval","author":"Zhu","year":"2022"},{"key":"10.1016\/j.patcog.2024.111225_bib0009","series-title":"Proceedings of the European Conference on Computer Vision (ECCV)","first-page":"202","article-title":"Cross-modal hamming hashing","author":"Cao","year":"2018"},{"key":"10.1016\/j.patcog.2024.111225_bib0010","series-title":"Proceedings of the 2019 on International Conference on Multimedia Retrieval","first-page":"159","article-title":"Adversary guided asymmetric hashing for cross-modal retrieval","author":"Gu","year":"2019"},{"key":"10.1016\/j.patcog.2024.111225_bib0011","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2023.110079","article-title":"Deep self-enhancement hashing for robust multi-label cross-modal retrieval","volume":"147","author":"Song","year":"2024","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.patcog.2024.111225_bib0012","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","article-title":"Pairwise relationship guided deep hashing for cross-modal retrieval","volume":"31","author":"Yang","year":"2017"},{"key":"10.1016\/j.patcog.2024.111225_bib0013","doi-asserted-by":"crossref","first-page":"304","DOI":"10.1016\/j.ins.2022.07.095","article-title":"Specific class center guided deep hashing for cross-modal retrieval","volume":"609","author":"Shu","year":"2022","journal-title":"Inf. Sci."},{"key":"10.1016\/j.patcog.2024.111225_bib0014","doi-asserted-by":"crossref","first-page":"138","DOI":"10.1016\/j.neucom.2021.09.053","article-title":"Multi-label enhancement based self-supervised deep cross-modal hashing","volume":"467","author":"Zou","year":"2022","journal-title":"Neurocomputing."},{"key":"10.1016\/j.patcog.2024.111225_bib0015","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"4242","article-title":"Self-supervised adversarial hashing networks for cross-modal retrieval","author":"Li","year":"2018"},{"issue":"12","key":"10.1016\/j.patcog.2024.111225_bib0016","doi-asserted-by":"crossref","first-page":"3101","DOI":"10.1109\/TMM.2020.2969792","article-title":"Multi-level correlation adversarial hashing for cross-modal retrieval","volume":"22","author":"Ma","year":"2020","journal-title":"IEEE Trans. Multimed."},{"key":"10.1016\/j.patcog.2024.111225_bib0017","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2021.107927","article-title":"Multi-label modality enhanced attention based self-supervised deep cross-modal hashing","volume":"239","author":"Zou","year":"2022","journal-title":"Knowl. Based Syst."},{"key":"10.1016\/j.patcog.2024.111225_bib0018","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2022.108676","article-title":"MS2GAH: multi-label semantic supervised graph attention hashing for robust cross-modal retrieval","volume":"128","author":"Duan","year":"2022","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.patcog.2024.111225_bib0019","series-title":"Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval","first-page":"686","article-title":"Data-aware proxy hashing for cross-modal retrieval","author":"Tu","year":"2023"},{"key":"10.1016\/j.patcog.2024.111225_bib0020","article-title":"Deep semantic-aware proxy hashing for multi-label cross-modal retrieval","author":"Huo","year":"2023","journal-title":"IEEE Trans. Circu. Syst. Video Technol."},{"key":"10.1016\/j.patcog.2024.111225_bib0021","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"4400","article-title":"Ranking-based deep cross-modal hashing","volume":"33","author":"Liu","year":"2019"},{"key":"10.1016\/j.patcog.2024.111225_bib0022","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"4941","article-title":"Deep normalized cross-modal hashing with bi-direction relation reasoning","author":"Sun","year":"2022"},{"key":"10.1016\/j.patcog.2024.111225_bib0023","series-title":"2020 IEEE International Conference on Multimedia and Expo (ICME)","first-page":"1","article-title":"Multi-similarity semantic correctional hashing for cross modal retrieval","author":"Zhan","year":"2020"},{"key":"10.1016\/j.patcog.2024.111225_bib0024","doi-asserted-by":"crossref","DOI":"10.1016\/j.image.2020.116131","article-title":"Multi-label semantics preserving based deep cross-modal hashing","volume":"93","author":"Zou","year":"2021","journal-title":"Signal Process. Image Commun."},{"key":"10.1016\/j.patcog.2024.111225_bib0025","series-title":"Proceedings of the European conference on Computer Vision (ECCV)","first-page":"3","article-title":"So Kweon, Cbam: convolutional block attention module","author":"Woo","year":"2018"},{"key":"10.1016\/j.patcog.2024.111225_bib0026","series-title":"Proceedings of the 2021 International Conference on Multimedia Retrieval","first-page":"376","article-title":"Teach: attention-aware deep cross-modal hashing","author":"Yao","year":"2021"},{"key":"10.1016\/j.patcog.2024.111225_bib0027","series-title":"Proceedings of the European Conference on Computer Vision (ECCV)","first-page":"591","article-title":"Attention-aware deep adversarial hashing for cross-modal retrieval","author":"Zhang","year":"2018"},{"issue":"4","key":"10.1016\/j.patcog.2024.111225_bib0028","doi-asserted-by":"crossref","first-page":"1519","DOI":"10.1007\/s11280-021-00881-8","article-title":"Deep medical cross-modal attention hashing","volume":"25","author":"Zhang","year":"2022","journal-title":"World Wide Web."},{"key":"10.1016\/j.patcog.2024.111225_bib0029","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"770","article-title":"Deep residual learning for image recognition","author":"He","year":"2016"},{"key":"10.1016\/j.patcog.2024.111225_bib0030","doi-asserted-by":"crossref","first-page":"400","DOI":"10.1214\/aoms\/1177729586","article-title":"A stochastic approximation method","author":"Robbins","year":"1951","journal-title":"Ann. Mathemat. Stat."},{"key":"10.1016\/j.patcog.2024.111225_bib0031","series-title":"Proceedings of the 1st ACM International Conference on Multimedia Information Retrieval","first-page":"39","article-title":"The MIR FLICKR retrieval evaluation","author":"Huiskes","year":"2008"},{"key":"10.1016\/j.patcog.2024.111225_bib0032","series-title":"Proceedings of the ACM International Conference on Image and Video Retrieval","first-page":"1","article-title":"Nus-wide: a real-world web image database from national university of singapore","author":"Chua","year":"2009"},{"issue":"4","key":"10.1016\/j.patcog.2024.111225_bib0033","doi-asserted-by":"crossref","first-page":"419","DOI":"10.1016\/j.cviu.2009.03.008","article-title":"The segmented and annotated IAPR TC-12 benchmark","volume":"114","author":"Escalante","year":"2010","journal-title":"Comput. Vis. Image Understand."},{"key":"10.1016\/j.patcog.2024.111225_bib0034","series-title":"European Conference on Computer Vision","first-page":"740","article-title":"Microsoft coco: common objects in context","author":"Lin","year":"2014"},{"key":"10.1016\/j.patcog.2024.111225_bib0035","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","article-title":"Large-scale supervised multimodal hashing with semantic correlation maximization","volume":"28","author":"Zhang","year":"2014"},{"key":"10.1016\/j.patcog.2024.111225_bib0036","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"3864","article-title":"Semantics-preserving hashing for cross-view retrieval","author":"Lin","year":"2015"},{"key":"10.1016\/j.patcog.2024.111225_bib0037","series-title":"IEEE Conference on Computer Vision and Pattern Recognition","first-page":"248","article-title":"Imagenet: a large-scale hierarchical image database","author":"Deng","year":"2009"},{"issue":"11","key":"10.1016\/j.patcog.2024.111225_bib0038","article-title":"Visualizing data using T-SNE","volume":"9","author":"Maaten","year":"2008","journal-title":"J. Mach. Learn. Res."}],"container-title":["Pattern Recognition"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0031320324009762?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0031320324009762?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2024,12,7]],"date-time":"2024-12-07T09:00:57Z","timestamp":1733562057000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0031320324009762"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4]]},"references-count":38,"alternative-id":["S0031320324009762"],"URL":"https:\/\/doi.org\/10.1016\/j.patcog.2024.111225","relation":{},"ISSN":["0031-3203"],"issn-type":[{"type":"print","value":"0031-3203"}],"subject":[],"published":{"date-parts":[[2025,4]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Semantic decomposition and enhancement hashing for deep cross-modal retrieval","name":"articletitle","label":"Article Title"},{"value":"Pattern Recognition","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.patcog.2024.111225","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2024 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"111225"}}