{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,8,7]],"date-time":"2024-08-07T01:13:42Z","timestamp":1722993222179},"reference-count":15,"publisher":"Fuji Technology Press Ltd.","issue":"3","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["J. Robot. Mechatron.","JRM"],"published-print":{"date-parts":[[2010,6,20]]},"abstract":"This paper presents a sound identification method for a mobile robot in home and office environments. We propose a short-term sound recognition method using Pitch-Cluster-Maps (PCMs) sound database (DB) based on a Vector Quantization approach. A binarized frequency spectrum is used to generate PCMs codebook, which describes a variety of sound sources, not only voice, from short-term sound input. PCMs sound identification requires several tens of milliseconds of sound input, and is suitable for mobile robot applications in which conditions are continuously and dynamically changing. We implemented this in mobile robot audition system using a 32-channel microphone array. Robot noise reduction and sound source tracking using our proposal are applied to robot audition system, and we evaluate daily sound recognition performance for separated sound sources from a moving robot.<\/jats:p>","DOI":"10.20965\/jrm.2010.p0402","type":"journal-article","created":{"date-parts":[[2016,4,14]],"date-time":"2016-04-14T02:23:03Z","timestamp":1460600583000},"page":"402-410","source":"Crossref","is-referenced-by-count":5,"title":["Pitch-Cluster-Map Based Daily Sound Recognition for Mobile Robot Audition"],"prefix":"10.20965","volume":"22","author":[{"given":"Yoko","family":"Sasaki","sequence":"first","affiliation":[]},{"name":"Digital Human Research Center, National Institute of Advanced Science and Technology, 2-3-26 Aomi, Kouto-ku, Tokyo 135-0064, Japan.","sequence":"first","affiliation":[]},{"given":"Masahito","family":"Kaneyoshi","sequence":"additional","affiliation":[]},{"given":"Satoshi","family":"Kagami","sequence":"additional","affiliation":[]},{"given":"Hiroshi","family":"Mizoguchi","sequence":"additional","affiliation":[]},{"given":"Tadashi","family":"Enomoto","sequence":"additional","affiliation":[]},{"name":"Dept. of Mechanical Engineering, Tokyo University of Science, 2641 Yamazaki, Noda-shi, Chiba 278-8510, Japan.","sequence":"additional","affiliation":[]},{"name":"The Kansai Electric Power Co. Inc., 3-11-20 Nakoji, Amagasaki, Hyogo 661-0974, Japan.","sequence":"additional","affiliation":[]}],"member":"8550","published-online":{"date-parts":[[2010,6,20]]},"reference":[{"unstructured":"S. Furui, \u201c50 years of progress in speech and speaker recognition,\u201d In Proc. of SPECOM2005, Patras, Greece, pp. 1-9, 2005.","key":"key-10.20965\/jrm.2010.p0402-1"},{"doi-asserted-by":"crossref","unstructured":"T. Matsui and K. Tanabe, \u201cComparative study of speaker identification methods : dplrm, svm and gmm,\u201d IEICE Trans. on INFOMATION and SYSTEMS, Vol.89-D, No.3, pp. 1066-1073, Mar., 2006.","key":"key-10.20965\/jrm.2010.p0402-2","DOI":"10.1093\/ietisy\/e89-d.3.1066"},{"doi-asserted-by":"crossref","unstructured":"N. Roman and D. L. Wang, \u201cPitch-based monaural segregation of reverberant speech,\u201d J. of Acoustics Society of America, Vol.120, No.1, pp. 458-469, Jul., 2006.","key":"key-10.20965\/jrm.2010.p0402-3","DOI":"10.1121\/1.2204590"},{"doi-asserted-by":"crossref","unstructured":"Y. Shao and D. L. Wang, \u201cModel-based sequential organization in cochannel speech,\u201d IEEE Trans. on Audio, Speech, and Language Processing, Vol.14, No.1, pp. 289-298, Jan., 2006.","key":"key-10.20965\/jrm.2010.p0402-4","DOI":"10.1109\/TSA.2005.854106"},{"unstructured":"M. Goto, \u201cAnalysis of musical audio signals. In D. L. Wang and G. J. Brown, editors, Computational Auditory Scene Analysis: Principles, Algorithms, and Applications,\u201d Wiley-IEEE Press, pp. 251-295. 2006.","key":"key-10.20965\/jrm.2010.p0402-5"},{"unstructured":"H. Fujihara, T. Kitahara, M. Goto, K. Komatani, T. Ogata, and H. G. Okuno, \u201cSinger identification based on accompaniment sound reduction and reliable frame selection,\u201d In Proc. of 6th Int. Conf. on Music Information Retrieval (ISMIR2005), London, U.K., pp. 329-336, Sep., 2005.","key":"key-10.20965\/jrm.2010.p0402-6"},{"unstructured":"H. Fujihara, T. Kitahara, M. Goto, K. Komatani, T. Ogata, and H. G. Okuno, \u201cSpeaker identification under noisy environments by using harmonic structure extraction and reliable frame weighting,\u201d In Proc. of Int. Conf. on Spoken Language Proc. (Interspeech2006), Pittsburgh PA, USA, pp. 1459-1462, Sep., 2006.","key":"key-10.20965\/jrm.2010.p0402-7"},{"unstructured":"J. Chen, A. H. Kam, J. Zhang, N. Liu, and L. Shue, \u201cBathroom activity monitoring based on sound,\u201d Pervasive Computing: Lecture notes in Computer Science, Vol.3468, pp. 47-61, May, 2005.","key":"key-10.20965\/jrm.2010.p0402-8"},{"unstructured":"K. Hiyane and J. Iio, \u201cNon-speech sound recognition with microphone array,\u201d In Proc. of IEEE Int. Workshop on Hands-Free Speech Communication (HSC2001), Kyoto, Japan, pp. 107-110, Apr., 2001.","key":"key-10.20965\/jrm.2010.p0402-9"},{"unstructured":"P. Lukowicz, J. AWard, H. Junker, M. Stager, G. Troster, A. Atrash, and T. Starner, \u201cRecognizing workshop activity using body worn microphones and accelerometers,\u201d Pervasive Computing: Lecture notes in Computer Science, Vol.3001, pp. 18-32, May, 2004.","key":"key-10.20965\/jrm.2010.p0402-10"},{"unstructured":"S. Tokutsu, K. Okada, and M. Inaba, \u201cDiscrimination of daily sounds for humanoids understanding situations,\u201d In Proc. of the 25th annual conf. of the Robotics Society of Japan, p. 1H36, Sep., 2007. (in japanese)","key":"key-10.20965\/jrm.2010.p0402-11"},{"unstructured":"Y. Sasaki, S. Kagami, and H. Mizoguchi, \u201cSimple sound source detection using main-lobe model of microphone array,\u201d In Proc. of the 25th annual conference of the Robotics Society of Japan, Chiba, Japan, p. 1N13, Sep., 2007. (in japanese)","key":"key-10.20965\/jrm.2010.p0402-12"},{"doi-asserted-by":"crossref","unstructured":"Y. Tamai, Y. Sasaki, S. Kagami, and H. Mizoguchi, \u201cThree ring microphone array for 3d sound localization and separation for mobile robot audition,\u201d In Proc. of 2005 IEEE\/RSJ Int. Conf. on Intelligent Robots and Systems (IROS2005), Edmonton, Canada, pp. 903-908, Aug., 2005.","key":"key-10.20965\/jrm.2010.p0402-13","DOI":"10.1109\/IROS.2005.1545095"},{"doi-asserted-by":"crossref","unstructured":"M. Aoki, M. Okamoto, S. Aoki, H. Matsui, T. Sakurai, and Y. Kaneda, \u201cSound source segregation based on estimating incident angle of each frequency component of input signals acquired by multiple microphones,\u201d Acoustical Science and Technology, Vol.22, No.2, pp. 149-157, 2001.","key":"key-10.20965\/jrm.2010.p0402-14","DOI":"10.1250\/ast.22.149"},{"doi-asserted-by":"crossref","unstructured":"T. Kinnunen and H. Li, \u201cAn overview of textindependent speaker recognition: From features to supervectors,\u201d Speech Communication, Vol.52, No.1, pp. 12-40, Jan. 2010.","key":"key-10.20965\/jrm.2010.p0402-15","DOI":"10.1016\/j.specom.2009.08.009"}],"container-title":["Journal of Robotics and Mechatronics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.fujipress.jp\/main\/wp-content\/themes\/Fujipress\/phyosetsu.php?ppno=ROBOT002200030019","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2017,6,25]],"date-time":"2017-06-25T02:59:22Z","timestamp":1498359562000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.fujipress.jp\/jrm\/rb\/robot002200030402"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2010,6,20]]},"references-count":15,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2010,6,20]]},"published-print":{"date-parts":[[2010,6,20]]}},"URL":"https:\/\/doi.org\/10.20965\/jrm.2010.p0402","relation":{},"ISSN":["1883-8049","0915-3942"],"issn-type":[{"type":"electronic","value":"1883-8049"},{"type":"print","value":"0915-3942"}],"subject":[],"published":{"date-parts":[[2010,6,20]]}}}