{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T03:17:35Z","timestamp":1740107855426,"version":"3.37.3"},"reference-count":28,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2023,6,6]],"date-time":"2023-06-06T00:00:00Z","timestamp":1686009600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,6,6]],"date-time":"2023-06-06T00:00:00Z","timestamp":1686009600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100007129","name":"Natural Science Foundation of Shandong Province","doi-asserted-by":"crossref","award":["ZR2020MF004","ZR2020MF004","ZR2020MF004","ZR2020MF004","ZR2020MF004","ZR2020MF004","ZR2020MF004","ZR2020MF004"],"id":[{"id":"10.13039\/501100007129","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100012166","name":"National Key R&D Program of China","doi-asserted-by":"crossref","award":["2020YFC0833201","2020YFC0833201","2020YFC0833201","2020YFC0833201","2020YFC0833201","2020YFC0833201","2020YFC0833201","2020YFC0833201"],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2023,8]]},"DOI":"10.1007\/s00530-023-01120-y","type":"journal-article","created":{"date-parts":[[2023,6,6]],"date-time":"2023-06-06T16:46:26Z","timestamp":1686069986000},"page":"2429-2437","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Micro-expression spotting network based on attention and one-dimensional convolutional sliding window"],"prefix":"10.1007","volume":"29","author":[{"given":"Hongbo","family":"Xing","sequence":"first","affiliation":[]},{"given":"Guanqun","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Shusen","family":"Yuan","sequence":"additional","affiliation":[]},{"given":"Youjun","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"Pinyong","family":"Geng","sequence":"additional","affiliation":[]},{"given":"Yewen","family":"Cao","sequence":"additional","affiliation":[]},{"given":"Yujun","family":"Li","sequence":"additional","affiliation":[]},{"given":"Lei","family":"Chen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,6,6]]},"reference":[{"key":"1120_CR1","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4684-6045-2_14","author":"EA Haggard","year":"1966","unstructured":"Haggard, E.A., Isaacs, K.S.: Micromomentary facial expressions as indicators of ego mechanisms in psychotherapy. Meth Res Psychoth. (1966). https:\/\/doi.org\/10.1007\/978-1-4684-6045-2_14","journal-title":"Meth Res Psychoth."},{"issue":"1","key":"1120_CR2","doi-asserted-by":"publisher","first-page":"88","DOI":"10.1080\/00332747.1969.11023575","volume":"32","author":"P Ekman","year":"1969","unstructured":"Ekman, P., Friesen, W.V.: Nonverbal leakage and clues to deception. Psychiatry 32(1), 88\u2013106 (1969). https:\/\/doi.org\/10.1080\/00332747.1969.11023575","journal-title":"Psychiatry"},{"key":"1120_CR3","doi-asserted-by":"publisher","first-page":"329","DOI":"10.3389\/fnins.2020.00329","volume":"14","author":"M Zhang","year":"2020","unstructured":"Zhang, M., Zhao, K., Qu, F., Li, K., Fu, X.: brain activation in contrasts of micro-expression following emotional contexts. Front. Neurosci. 14, 329\u2013329 (2020). https:\/\/doi.org\/10.3389\/fnins.2020.00329","journal-title":"Front. Neurosci."},{"issue":"6","key":"1120_CR4","doi-asserted-by":"publisher","first-page":"802","DOI":"10.1080\/10503307.2020.1836424","volume":"31","author":"DA Curtis","year":"2021","unstructured":"Curtis, D.A.: Deception detection and emotion recognition: Investigating FACE software. Psychother Res. 31(6), 802\u2013816 (2021). https:\/\/doi.org\/10.1080\/10503307.2020.1836424","journal-title":"Psychother Res."},{"issue":"24","key":"1120_CR5","doi-asserted-by":"publisher","first-page":"4901","DOI":"10.3390\/ijerph16244901","volume":"16","author":"F Datz","year":"2019","unstructured":"Datz, F., Wong, G., L\u00f6ffler-Stastka, H.: Interpretation and Working through Contemptuous Facial Micro-Expressions Benefits the Patient-Therapist Relationship. Int J Environ Res Public Health 16(24), 4901 (2019). https:\/\/doi.org\/10.3390\/ijerph16244901","journal-title":"Int J Environ Res Public Health"},{"issue":"4","key":"1120_CR6","doi-asserted-by":"publisher","first-page":"622","DOI":"10.1057\/s41284-019-00204-7","volume":"33","author":"LM Jupe","year":"2019","unstructured":"Jupe, L.M., Keatley, D.A.: Airport artificial intelligence can detect deception: or am i lying? Secur. J. 33(4), 622\u2013635 (2019). https:\/\/doi.org\/10.1057\/s41284-019-00204-7","journal-title":"Secur. J."},{"issue":"4","key":"1120_CR7","doi-asserted-by":"publisher","first-page":"424","DOI":"10.1109\/taffc.2017.2654440","volume":"9","author":"F Qu","year":"2018","unstructured":"Qu, F., Wang, S.-J., Yan, W.-J., Li, H., Wu, S., Fu, X.: CAS(ME)^2: A Database for Spontaneous Macro-expression and Micro-expression Spotting and Recognition. IEEE Trans. Affect. Comput. 9(4), 424\u2013436 (2018). https:\/\/doi.org\/10.1109\/taffc.2017.2654440","journal-title":"IEEE Trans. Affect. Comput."},{"key":"1120_CR8","doi-asserted-by":"publisher","first-page":"356","DOI":"10.1016\/j.neucom.2021.02.022","volume":"443","author":"T Tran","year":"2021","unstructured":"Tran, T., Vo, Q., Hong, X., Li, X., Zhao, G.: Micro-expression spotting: A new benchmark. Neurocomputing (Amsterdam) 443, 356\u2013368 (2021). https:\/\/doi.org\/10.1016\/j.neucom.2021.02.022","journal-title":"Neurocomputing (Amsterdam)"},{"key":"1120_CR9","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR.2014.303","author":"A Moilanen","year":"2014","unstructured":"Moilanen, A., Zhao, G., Pietikainen, M.: Spotting rapid facial movements from videos using appearance-based feature difference analysis. Int Conf Pattern Recog (2014). https:\/\/doi.org\/10.1109\/ICPR.2014.303","journal-title":"Int Conf Pattern Recog"},{"key":"1120_CR10","doi-asserted-by":"publisher","DOI":"10.1109\/smc.2015.326","author":"AK Davison","year":"2015","unstructured":"Davison, A.K., Yap, M.H., Lansley, C.: Micro-facial movement detection using individualised baselines and histogram-based descriptors. Int Conf Syst and Cybernet. (2015). https:\/\/doi.org\/10.1109\/smc.2015.326","journal-title":"Int Conf Syst and Cybernet."},{"issue":"4","key":"1120_CR11","doi-asserted-by":"publisher","first-page":"563","DOI":"10.1109\/TAFFC.2017.2667642","volume":"9","author":"X Li","year":"2017","unstructured":"Li, X., Hong, X., Moilanen, A., Huang, X., Pfister, T., Zhao, G., Pietikainen, M.: Towards reading hidden emotions: a comparative study of spontaneous micro-expression spotting and recognition methods. Trans Affect Comp. 9(4), 563\u2013577 (2017). https:\/\/doi.org\/10.1109\/TAFFC.2017.2667642","journal-title":"Trans Affect Comp."},{"key":"1120_CR12","doi-asserted-by":"publisher","first-page":"382","DOI":"10.1016\/j.neucom.2016.12.034","volume":"230","author":"S-J Wang","year":"2017","unstructured":"Wang, S.-J., Wu, S., Qian, X., Li, J., Fu, X.: A main directional maximal difference analysis for spotting facial movements from long-term videos. Neurocomputing 230, 382\u2013389 (2017). https:\/\/doi.org\/10.1016\/j.neucom.2016.12.034","journal-title":"Neurocomputing"},{"key":"1120_CR13","doi-asserted-by":"publisher","DOI":"10.1109\/CSPA.2019.8696059","author":"KX Beh","year":"2019","unstructured":"Beh, K.X., Goh, K.M.: Micro-Expression spotting using facial landmarks. In CSPA. (2019). https:\/\/doi.org\/10.1109\/CSPA.2019.8696059","journal-title":"In CSPA."},{"key":"1120_CR14","doi-asserted-by":"publisher","first-page":"57","DOI":"10.1016\/j.patrec.2022.09.009","volume":"163","author":"Y He","year":"2022","unstructured":"He, Y., Xu, Z., Ma, L., Li, H.: Micro-expression spotting based on optical flow features. Patt recog lett 163, 57\u201363 (2022)","journal-title":"Patt recog lett"},{"key":"1120_CR15","doi-asserted-by":"publisher","first-page":"87","DOI":"10.1016\/j.cviu.2015.12.006","volume":"147","author":"Z Xia","year":"2016","unstructured":"Xia, Z., Feng, X., Peng, J., Peng, X., Zhao, G.: Spontaneous micro-expression spotting via geometric deformation modeling. Comput Vis Image Underst. 147, 87\u201394 (2016). https:\/\/doi.org\/10.1016\/j.cviu.2015.12.006","journal-title":"Comput Vis Image Underst."},{"key":"1120_CR16","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-70353-4_46","author":"TK Tran","year":"2017","unstructured":"Tran, T.K., Hong, X., Zhao, G.: Sliding Window Based Micro-expression Spotting: A Benchmark. Adv Conc Intell Vision Syst. (2017). https:\/\/doi.org\/10.1007\/978-3-319-70353-4_46","journal-title":"Adv Conc Intell Vision Syst."},{"key":"1120_CR17","first-page":"7113","volume":"6","author":"Z Zhang","year":"2018","unstructured":"Zhang, Z., Chen, T., Meng, H., Liu, G., Fu, X.: Smeconvnet a convolutional neural network for spotting spontaneous facial micro-expression from long videos. IEEE Access. 6, 7113 (2018)","journal-title":"IEEE Access."},{"key":"1120_CR18","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2019.8756588","author":"M Verburg","year":"2019","unstructured":"Verburg, M., Menkovski, V.: Micro-expression detection in long videos using optical flow and recurrent neural networks. In Fg (2019). https:\/\/doi.org\/10.1109\/FG.2019.8756588","journal-title":"In Fg"},{"key":"1120_CR19","doi-asserted-by":"publisher","first-page":"401","DOI":"10.2352\/issn.2470-1173.2019.8.imawm-401","volume":"8","author":"T-K Tran","year":"2019","unstructured":"Tran, T.-K., Vo, Q.-N., Hong, X., Zhao, G.: Dense prediction for micro-expression spotting based on deep sequence model. Electr Imag. 8, 401\u2013411 (2019). https:\/\/doi.org\/10.2352\/issn.2470-1173.2019.8.imawm-401","journal-title":"Electr Imag."},{"key":"1120_CR20","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00041","author":"CR Chen","year":"2021","unstructured":"Chen, C.R., Fan, Q., Panda, R.: CrossViT: cross-attention multi-scale vision transformer for image classification. ICCV. (2021). https:\/\/doi.org\/10.1109\/ICCV48922.2021.00041","journal-title":"ICCV."},{"key":"1120_CR21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01173","author":"X Mao","year":"2022","unstructured":"Mao, X., Qi, G., Chen, Y., et al.: Towards robust vision transformer. EEE\/CVF Conf Comp Vis Patt Recog. (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.01173","journal-title":"EEE\/CVF Conf Comp Vis Patt Recog."},{"key":"1120_CR22","doi-asserted-by":"publisher","first-page":"104378","DOI":"10.1016\/j.imavis.2022.104378","volume":"119","author":"Y Zhou","year":"2022","unstructured":"Zhou, Y., Song, Y., Chen, L., Chen, Y., Ben, X., Cao, Y.: A novel micro-expression detection algorithm based on BERT and 3DCNN. Image Vision Comp 119, 104378 (2022). https:\/\/doi.org\/10.1016\/j.imavis.2022.104378","journal-title":"Image Vision Comp"},{"issue":"2","key":"1120_CR23","doi-asserted-by":"publisher","first-page":"1489","DOI":"10.1109\/TPAMI.2022.3164083","volume":"45","author":"Y Li","year":"2023","unstructured":"Li, Y., Yao, T., Pan, Y., Mei, T.: Contextual transformer networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 45(2), 1489\u20131500 (2023). https:\/\/doi.org\/10.1109\/TPAMI.2022.3164083","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1120_CR24","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2023.3268446","author":"T Yao","year":"2023","unstructured":"Yao, T., Li, Y., Pan, Y., Wang, Y., Zhang, X., Mei, T.: Dual vision transformer. IEEE Trans. Pattern Anal. Mach. Intell. (2023). https:\/\/doi.org\/10.1109\/TPAMI.2023.3268446","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"6","key":"1120_CR25","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2017","unstructured":"Krizhevsky, A., Sutskeve, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. Commun. ACM 60(6), 84\u201390 (2017). https:\/\/doi.org\/10.1145\/3065386","journal-title":"Commun. ACM"},{"key":"1120_CR26","doi-asserted-by":"publisher","first-page":"824592","DOI":"10.3389\/fnbot.2021.824592","volume":"15","author":"C Wang","year":"2022","unstructured":"Wang, C., Wang, Z.: Progressive Multi-Scale Vision Transformer for Facial Action Unit Detection. Front Neurorobot. 15, 824592 (2022). https:\/\/doi.org\/10.3389\/fnbot.2021.824592","journal-title":"Front Neurorobot."},{"key":"1120_CR27","first-page":"1755","volume":"10","author":"DE King","year":"2009","unstructured":"King, D.E.: Dlib-ml: a machine learning toolkit. J Mach Learn Res 10, 1755\u20131758 (2009)","journal-title":"J Mach Learn Res"},{"key":"1120_CR28","doi-asserted-by":"publisher","DOI":"10.1007\/3-540-45103-x_50","author":"G Farneb\u00e4ck","year":"2003","unstructured":"Farneb\u00e4ck, G.: Two-frame motion estimation based on polynomial expansion. Lect Not Comp Sci (2003). https:\/\/doi.org\/10.1007\/3-540-45103-x_50","journal-title":"Lect Not Comp Sci"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-023-01120-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-023-01120-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-023-01120-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,7,14]],"date-time":"2023-07-14T10:33:16Z","timestamp":1689330796000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-023-01120-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,6]]},"references-count":28,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2023,8]]}},"alternative-id":["1120"],"URL":"https:\/\/doi.org\/10.1007\/s00530-023-01120-y","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"type":"print","value":"0942-4962"},{"type":"electronic","value":"1432-1882"}],"subject":[],"published":{"date-parts":[[2023,6,6]]},"assertion":[{"value":"6 January 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 May 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 June 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Authors declare that there is no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}