{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,23]],"date-time":"2024-10-23T03:40:50Z","timestamp":1729654850685,"version":"3.28.0"},"reference-count":61,"publisher":"IEEE","license":[{"start":{"date-parts":[[2021,10,1]],"date-time":"2021-10-01T00:00:00Z","timestamp":1633046400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2021,10,1]],"date-time":"2021-10-01T00:00:00Z","timestamp":1633046400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012492","name":"Youth Innovation Promotion Association","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100012492","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,10]]},"DOI":"10.1109\/iccv48922.2021.01090","type":"proceedings-article","created":{"date-parts":[[2022,2,28]],"date-time":"2022-02-28T22:08:02Z","timestamp":1646086082000},"page":"11067-11076","source":"Crossref","is-referenced-by-count":39,"title":["SignBERT: Pre-Training of Hand-Model-Aware Representation for Sign Language Recognition"],"prefix":"10.1109","author":[{"given":"Hezhen","family":"Hu","sequence":"first","affiliation":[{"name":"University of Science and Technology of China (USTC),CAS Key Laboratory of GIPAS,EEIS Department"}]},{"given":"Weichao","family":"Zhao","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China (USTC),CAS Key Laboratory of GIPAS,EEIS Department"}]},{"given":"Wengang","family":"Zhou","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China (USTC),CAS Key Laboratory of GIPAS,EEIS Department"}]},{"given":"Yuechen","family":"Wang","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China (USTC),CAS Key Laboratory of GIPAS,EEIS Department"}]},{"given":"Houqiang","family":"Li","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China (USTC),CAS Key Laboratory of GIPAS,EEIS Department"}]}],"member":"263","reference":[{"key":"ref39","first-page":"1","article-title":"PyTorch: An imperative style, high-performance deep learning library","author":"paszke","year":"2019","journal-title":"NeurIPS"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.437"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00624"},{"journal-title":"Co-occurrence feature learning from skeleton data for action recognition and detection with hierarchical aggregation","first-page":"786","year":"2018","author":"li","key":"ref32"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1145\/344779.344862"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-005-1838-7"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00580"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00718"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i07.6795"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/WACV45572.2020.9093512"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00877"},{"key":"ref61","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.525"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2019.2911077"},{"first-page":"1","article-title":"Quantitative survey of the state of the art in sign language recognition","year":"2020","author":"koller","key":"ref27"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-018-1121-3"},{"key":"ref2","first-page":"640","article-title":"Motion capture of hands in action using discriminative salient points","author":"ballan","year":"2012","journal-title":"ECCV"},{"key":"ref1","first-page":"35","article-title":"BSL-1k: Scaling up co-articulated sign language recognition using mouthing cues","author":"albanie","year":"2020","journal-title":"ECCV"},{"key":"ref20","doi-asserted-by":"crossref","first-page":"1558","DOI":"10.1609\/aaai.v35i2.16247","article-title":"Hand-model-aware sign language recognition","author":"hu","year":"2021","journal-title":"AAAI"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2018.2870740"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1145\/3436754"},{"key":"ref24","first-page":"1","article-title":"MS-ASL: A largescale data set and benchmark for understanding american sign language","author":"vaezi joze","year":"2019","journal-title":"BMVC"},{"key":"ref23","first-page":"2257","article-title":"Video-based sign language recognition without temporal segmentation","author":"huang","year":"2018","journal-title":"AAAI"},{"key":"ref26","first-page":"3294","article-title":"Skip-thought vectors","author":"kiros","year":"2015","journal-title":"NeurIPS"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1145\/1053427.1053429"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1145\/2980179.2980226"},{"key":"ref51","first-page":"31","article-title":"Pose-based sign language recognition using GCN and BERT","author":"tunga","year":"2020","journal-title":"Workshop on WACV"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00137"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2017.8296428"},{"first-page":"1","article-title":"The 2017 hands in the million challenge on 3D hand pose estimation","year":"2017","author":"yuan","key":"ref57"},{"key":"ref56","first-page":"1","article-title":"XLNet: Generalized autoregressive pretraining for language understanding","author":"yang","year":"2019","journal-title":"NeurIPS"},{"key":"ref55","first-page":"7444","article-title":"Spatial temporal graph convolutional networks for skeleton-based action recognition","author":"yan","year":"2018","journal-title":"AAAI"},{"first-page":"1","article-title":"Google's neural machine translation system: Bridging the gap between human and machine translation","year":"2016","author":"wu","key":"ref54"},{"key":"ref53","first-page":"5999","article-title":"Attention is all you need","author":"vaswani","year":"2017","journal-title":"NeurIPS"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-016-0895-4"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58586-0_41"},{"journal-title":"Openmmlab pose estimation toolbox and benchmark","year":"2020","key":"ref11"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1162"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2018.2889563"},{"key":"ref13","first-page":"248","article-title":"ImageNet: A large-scale hierarchical image database","author":"deng","year":"2009","journal-title":"CVPR"},{"key":"ref14","first-page":"4171","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","author":"devlin","year":"2018","journal-title":"NAACL"},{"key":"ref15","first-page":"1110","article-title":"Hierarchical recurrent neural network for skeleton based action recognition","author":"du","year":"2015","journal-title":"CVPR"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58555-6_40"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00630"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00510"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01616"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00236"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.01110"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01004"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.332"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.502"},{"key":"ref7","first-page":"3247","article-title":"Skeleton-based action recognition with gated convolutional neural networks","volume":"29","author":"cao","year":"2018","journal-title":"TCSVT"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1145\/2735952"},{"key":"ref9","first-page":"1691","article-title":"Generative pretraining from pixels","author":"chen","year":"2020","journal-title":"ICML"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.305"},{"key":"ref45","first-page":"4263","article-title":"An end-to-end spatio-temporal attention model for human action recognition from skeleton data","author":"song","year":"2017","journal-title":"AAAI"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00756"},{"key":"ref47","first-page":"1","article-title":"VL-BERT: Pre-training of generic visual-linguistic representations","author":"su","year":"2020","journal-title":"ICLRE"},{"key":"ref42","first-page":"5533","article-title":"Learning spatio-temporal representation with pseudo-3D residual networks","author":"qiu","year":"2017","journal-title":"ICCV"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.145"},{"key":"ref44","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3130800.3130883","article-title":"Embodied hands: Modeling and capturing hands and bodies together","volume":"36","author":"romero","year":"2017","journal-title":"TOG"},{"first-page":"1","article-title":"Improving language understanding by generative pre-training","year":"2018","author":"radford","key":"ref43"}],"event":{"name":"2021 IEEE\/CVF International Conference on Computer Vision (ICCV)","start":{"date-parts":[[2021,10,10]]},"location":"Montreal, QC, Canada","end":{"date-parts":[[2021,10,17]]}},"container-title":["2021 IEEE\/CVF International Conference on Computer Vision (ICCV)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9709627\/9709628\/09709967.pdf?arnumber=9709967","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,1,28]],"date-time":"2023-01-28T03:06:53Z","timestamp":1674875213000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9709967\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,10]]},"references-count":61,"URL":"https:\/\/doi.org\/10.1109\/iccv48922.2021.01090","relation":{},"subject":[],"published":{"date-parts":[[2021,10]]}}}