{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,23]],"date-time":"2024-10-23T10:15:03Z","timestamp":1729678503245,"version":"3.28.0"},"reference-count":27,"publisher":"IEEE","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2015,9]]},"DOI":"10.1109\/acii.2015.7344606","type":"proceedings-article","created":{"date-parts":[[2015,12,8]],"date-time":"2015-12-08T22:20:36Z","timestamp":1449613236000},"page":"428-434","source":"Crossref","is-referenced-by-count":3,"title":["GMM-based synchronization rules for HMM-based audio-visual laughter synthesis"],"prefix":"10.1109","author":[{"given":"Huseyin","family":"Cakmak","sequence":"first","affiliation":[]},{"given":"Kevin El","family":"Haddad","sequence":"additional","affiliation":[]},{"given":"Thierry","family":"Dutoit","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2014.6854469"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/JSTSP.2013.2281036"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.1998.679698"},{"key":"ref13","article-title":"Visual speech synthesis based on parameter generation from hmm: Speech-driven and text-and-speech-driven approaches","author":"tamura","year":"1998","journal-title":"Int Conf Auditory-Visual Speech Process"},{"key":"ref14","article-title":"Hmm-based synthesis of laughter facial expression","author":"\u00e7akmak","year":"2015","journal-title":"Journal on Multimodal User Interfaces"},{"key":"ref15","article-title":"Synchronization rules for HMM-based audio-visual laughter synthesis","author":"h\u00fcseyin","year":"2015","journal-title":"IEEE International Conference on Acoustics Speech and Signal Processing ICASSP"},{"key":"ref16","doi-asserted-by":"crossref","DOI":"10.1002\/0471721182","author":"mclachlan","year":"0","journal-title":"Finite Mixture Models 2000"},{"key":"ref17","article-title":"The av-lasyn database: A synchronous corpus of audio and 3d facial marker data for audio-visual laughter synthesis","author":"\u00e7akmak","year":"2014","journal-title":"Proc of the 9th Int Conf on Language Resources and Evaluation (LREC‘14)"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-24600-5_43"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1142\/9789812810687_0033"},{"key":"ref4","doi-asserted-by":"crossref","DOI":"10.21437\/Interspeech.2008-591","article-title":"Speech-driven lip motion generation with a trajectory hmm","author":"hofer","year":"2008"},{"key":"ref27","doi-asserted-by":"crossref","first-page":"593","DOI":"10.21437\/Interspeech.2011-239","article-title":"Statistical mapping between articulatory and acoustic data for an ultrasound-based silent speech interface","author":"hueber","year":"2011","journal-title":"InterSpeech"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/JSTSP.2014.2309435"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2011.5947374"},{"key":"ref5","article-title":"Hmm-based text-to-audio-visual speech synthesis","author":"sako","year":"0","journal-title":"ICSLP - 2000"},{"key":"ref8","article-title":"Learning optimal audiovisual phasing for a hmm-based control model for facial animation","author":"govokhina","year":"2007","journal-title":"6th ISCA Workshop on Speech (SSW6)"},{"article-title":"Comparison of hmm and tmd methods for lip synchronisation","year":"2010","author":"hofer","key":"ref7"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/ACII.2013.32"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1155\/2009\/769494"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2013.6639189"},{"key":"ref20","article-title":"The hmm-based speech synthesis system (hts) version 2.0","author":"zen","year":"0","journal-title":"Proc of Sixth ISCA Workshop on Speech 2007"},{"journal-title":"Applied Smoothing Techniques for Data Analysis The Kernel Approach with S-Plus Illustrations The Kernel Approach with S-Plus Illustrations","year":"1997","author":"bowman","key":"ref22"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1142\/9789812810687_0033"},{"key":"ref24","article-title":"A method for silence removal and segmentation of speech signals, implemented in matlab","author":"giannakopoulos","year":"0","journal-title":"University of Athens Athens 2009"},{"key":"ref23","article-title":"Continuous univariate distributions","volume":"2","author":"johnson","year":"1995","journal-title":"in Wiley series in probability and mathematical statistics Applied probability and statistics Wiley & Sons"},{"journal-title":"High resolution voice transformation","year":"2001","author":"blouke kain","key":"ref26"},{"journal-title":"Harmonic plus noise models for speech combined with statistical methods for speech and speaker modification","year":"1996","author":"stylianou","key":"ref25"}],"event":{"name":"2015 International Conference on Affective Computing and Intelligent Interaction (ACII)","start":{"date-parts":[[2015,9,21]]},"location":"Xi'an, China","end":{"date-parts":[[2015,9,24]]}},"container-title":["2015 International Conference on Affective Computing and Intelligent Interaction (ACII)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/7332343\/7344542\/07344606.pdf?arnumber=7344606","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,28]],"date-time":"2022-05-28T15:32:57Z","timestamp":1653751977000},"score":1,"resource":{"primary":{"URL":"http:\/\/ieeexplore.ieee.org\/document\/7344606\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2015,9]]},"references-count":27,"URL":"https:\/\/doi.org\/10.1109\/acii.2015.7344606","relation":{},"subject":[],"published":{"date-parts":[[2015,9]]}}}