{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T01:23:25Z","timestamp":1740101005632,"version":"3.37.3"},"reference-count":45,"publisher":"IEEE","license":[{"start":{"date-parts":[[2022,7,18]],"date-time":"2022-07-18T00:00:00Z","timestamp":1658102400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,7,18]],"date-time":"2022-07-18T00:00:00Z","timestamp":1658102400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"NSFC","doi-asserted-by":"publisher","award":["62076096,62006078"],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003399","name":"STCSM","doi-asserted-by":"publisher","award":["22ZR1421700"],"id":[{"id":"10.13039\/501100003399","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022,7,18]]},"DOI":"10.1109\/ijcnn55064.2022.9892637","type":"proceedings-article","created":{"date-parts":[[2022,9,30]],"date-time":"2022-09-30T19:56:04Z","timestamp":1664567764000},"page":"1-8","source":"Crossref","is-referenced-by-count":1,"title":["Robust Cross-Modal Retrieval by Adversarial Training"],"prefix":"10.1109","author":[{"given":"Tao","family":"Zhang","sequence":"first","affiliation":[{"name":"School of Computer Science and Technology, East China Normal University,Shanghai,China"}]},{"given":"Shiliang","family":"Sun","sequence":"additional","affiliation":[{"name":"School of Computer Science and Technology, East China Normal University,Shanghai,China"}]},{"given":"Jing","family":"Zhao","sequence":"additional","affiliation":[{"name":"School of Computer Science and Technology, East China Normal University,Shanghai,China"}]}],"member":"263","reference":[{"key":"ref39","first-page":"1","article-title":"Contrastive learning with adversarial examples","author":"ho","year":"0","journal-title":"Proceedings of Annual Conference on Neural Information Processing Systems"},{"key":"ref38","first-page":"1","article-title":"Adversarial self-supervised contrastive learning","author":"kim","year":"0","journal-title":"Proceedings of Annual Conference on Neural Information Processing Systems"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1670"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1145\/3308558.3313416"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.361"},{"key":"ref30","first-page":"5858","article-title":"Adversarial training and robustness for multiple perturbations","author":"tram\u00e8r","year":"0","journal-title":"Proceedings of Annual Conference on Neural Information Processing Systems"},{"key":"ref37","first-page":"1","article-title":"Robust pre-training by adversarial contrastive learning","author":"jiang","year":"0","journal-title":"Proceedings of Advances in Neural Information Processing Systems"},{"key":"ref36","first-page":"1597","article-title":"A simple frame-work for contrastive learning of visual representations","volume":"119","author":"chen","year":"0","journal-title":"Proceedings of the 37th International Conference on Machine Learning"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1425"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.500"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i07.6708"},{"key":"ref40","first-page":"1074","article-title":"Adco: Adversarial contrast for efficient learning of unsupervised representations from self-trained negative adversaries","author":"hu","year":"0","journal-title":"Proceedings of IEEE Conference on Computer Vision and Pattern Recognition"},{"key":"ref11","first-page":"10 791","article-title":"Cross-modal learning with adversarial samples","author":"li","year":"0","journal-title":"Proceedings of Advances in Neural Information Processing Systems"},{"key":"ref12","article-title":"Large-scale adversarial training for vision-and-language representation learning","author":"gan","year":"0","journal-title":"Proceedings of Advances in Neural Information Processing Systems"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1145\/3440755"},{"key":"ref14","first-page":"1","article-title":"A comprehensive survey on cross-modal retrieval","volume":"abs 1607 6215","author":"wang","year":"2016","journal-title":"CoRR"},{"key":"ref15","first-page":"251","article-title":"A new approach to cross-modal multi-media retrieval","author":"rasiwasia","year":"0","journal-title":"Proc of the 18th International Conf on Multimedia"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1007\/11752790_2"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2012.6247923"},{"key":"ref18","first-page":"3270","article-title":"Deep cross-modal hashing","author":"jiang","year":"0","journal-title":"Proceedings of IEEE Conference on Computer Vision and Pattern Recognition"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v31i1.10719"},{"key":"ref28","first-page":"2574","article-title":"Deepfool: A simple and accurate method to fool deep neural networks","author":"moosavi-dezfooli","year":"0","journal-title":"Proceedings of IEEE Conference on Computer Vision and Pattern Recognition"},{"key":"ref4","first-page":"6345","article-title":"Cross-modality binary code learning via fusion similarity hashing","author":"liu","year":"0","journal-title":"Proceedings of IEEE Conference on Computer Vision and Pattern Recognition"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2017.49"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00855"},{"key":"ref6","first-page":"2940","article-title":"Learning aligned cross-modal representations from weakly aligned data","author":"castrej\u00f3n","year":"0","journal-title":"Proceedings of IEEE Conference on Computer Vision and Pattern Recognition"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58452-8_36"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2018.2798607"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58577-8_7"},{"key":"ref7","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume":"139","author":"radford","year":"0","journal-title":"Proceedings of the 38th International Conference on Machine Learning"},{"key":"ref2","first-page":"1060","article-title":"Generative adversarial text to image synthesis","volume":"48","author":"reed","year":"0","journal-title":"Proceedings of the 33nd International Conference on Machine Learning"},{"key":"ref9","first-page":"1","article-title":"Wenlan: Bridging vision and language by large-scale multi-modal pre-training","volume":"abs 2103 6561","author":"huo","year":"2021","journal-title":"CoRR"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1145\/3123266.3123326"},{"key":"ref20","first-page":"4242","article-title":"Self-supervised adversarial hashing networks for cross-modal retrieval","author":"li","year":"0","journal-title":"Proceedings of IEEE Conference on Computer Vision and Pattern Recognition"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3403084"},{"key":"ref22","first-page":"14 558","article-title":"MCEN: bridging cross-modal gap between cooking recipes and dish images with latent variable model","author":"fu","year":"0","journal-title":"Proceedings of IEEE Conference on Computer Vision and Pattern Recognition"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01246-5_13"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1145\/1460096.1460104"},{"key":"ref24","first-page":"1","article-title":"Intriguing properties of neural networks","author":"szegedy","year":"0","journal-title":"Proceedings of the 2nd International Conference on Learning Representations"},{"key":"ref41","first-page":"740","article-title":"Microsoft COCO: common objects in context","volume":"8693","author":"lin","year":"0","journal-title":"Proceedings of the 13th European Conference on Computer Vision"},{"key":"ref23","first-page":"707","article-title":"Deep cross-modal projection learning for image-text matching","volume":"11205","author":"zhang","year":"0","journal-title":"Proceedings of the 15th European Conference on Computer Vision"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.5244\/C.28.6"},{"key":"ref26","first-page":"1","article-title":"Towards deep learning models resistant to adversarial attacks","author":"madry","year":"0","journal-title":"Proceedings of the 6th International Conference on Learning Representations"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.598"},{"key":"ref25","first-page":"1","article-title":"Explaining and harnessing adversarial examples","author":"goodfellow","year":"0","journal-title":"Proceedings of the 3rd International Conference on Learning Representations"}],"event":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","start":{"date-parts":[[2022,7,18]]},"location":"Padua, Italy","end":{"date-parts":[[2022,7,23]]}},"container-title":["2022 International Joint Conference on Neural Networks (IJCNN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9891857\/9889787\/09892637.pdf?arnumber=9892637","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,11,3]],"date-time":"2022-11-03T23:00:08Z","timestamp":1667516408000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9892637\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,7,18]]},"references-count":45,"URL":"https:\/\/doi.org\/10.1109\/ijcnn55064.2022.9892637","relation":{},"subject":[],"published":{"date-parts":[[2022,7,18]]}}}