default search action
Computer Speech & Language, Volume 27
Volume 27, Number 1, January 2013
- Björn W. Schuller, Stefan Steidl, Anton Batliner:
Introduction to the special issue on Paralinguistics in Naturalistic Speech and Language. 1-3
- Björn W. Schuller, Stefan Steidl, Anton Batliner, Felix Burkhardt, Laurence Devillers, Christian A. Müller, Shrikanth S. Narayanan:
Paralinguistics in speech and language - State-of-the-art and the challenge. 4-39 - Klaus R. Scherer:
Vocal markers of emotion: Comparing induction and acting elicitation. 40-58 - Abualsoud Hanani, Martin J. Russell, Michael J. Carey:
Human and computer recognition of regional accents and ethnic groups from British English speech. 59-74 - Luís Pinto Coelho, Daniela Braga, Miguel Sales Dias, Carmen García-Mateo:
On the development of an automatic voice pleasantness classification and intensity estimation system. 75-88 - Rajesh Ranganath, Dan Jurafsky, Daniel A. McFarland:
Detecting friendly, flirtatious, awkward, and assertive speech in speed-dates. 89-115 - Serguei V. S. Pakhomov, Susan E. Marino, Angela K. Birnbaum:
Quantification of speech disfluency as a marker of medication-induced cognitive impairment: An application of computerized speech analysis in neuropharmacology. 116-134 - Rok Gajsek, France Mihelic, Simon Dobrisek:
Speaker state recognition using an HMM-based feature extraction method. 135-150 - Ming Li, Kyu Jeong Han, Shrikanth S. Narayanan:
Automatic speaker age and gender recognition using acoustic and prosodic level information fusion. 151-167 - William Yang Wang, Fadi Biadsy, Andrew Rosenberg, Julia Hirschberg:
Automatic detection of speaker state: Lexical, prosodic, and phonetic approaches to level-of-interest and intoxication classification. 168-189
- Antonio Origlia, Giovanni Abete, Francesco Cutugno:
A dynamic tonal perception model for optimal pitch stylization. 190-208 - Sabato Marco Siniscalchi, Jeremy Reed, Torbjørn Svendsen, Chin-Hui Lee:
Universal attribute characterization of spoken languages for automatic spoken language recognition. 209-227 - Izabela Swietlicka, Wieslawa Kuniszyk-Józkowiak, Elzbieta Smolka:
Hierarchical ANN system for stuttering identification. 228-242 - Gabriel Skantze, Anna Hjalmarsson:
Towards incremental speech generation in conversational systems. 243-262 - Stefan Scherer, John Kane, Christer Gobl, Friedhelm Schwenker:
Investigating fuzzy-input fuzzy-output support vector machines for robust voice quality classification. 263-287 - Panikos Heracleous, Carlos Toshinori Ishi, Miki Sato, Hiroshi Ishiguro, Norihiro Hagita:
Analysis of the visual Lombard effect and automatic recognition experiments. 288-300 - Xunying Liu, Mark John Francis Gales, Philip C. Woodland:
Use of contexts in language model interpolation and adaptation. 301-321 - Rogier C. van Dalen, Mark John Francis Gales:
Importance sampling to compute likelihoods of noise-corrupted speech. 322-349 - Marc Delcroix, Shinji Watanabe, Tomohiro Nakatani, Atsushi Nakamura:
Cluster-based dynamic variance adaptation for interconnecting speech enhancement pre-processor and speech recognizer. 350-368 - Satoshi Kobashikawa, Atsunori Ogawa, Taichi Asami, Yoshikazu Yamaguchi, Hirokazu Masataki, Satoshi Takahashi:
Fast unsupervised adaptation based on efficient statistics accumulation using frame independent confidence within monophone states. 369-379 - Alexandros Tsilfidis, Iosif Mporas, John Mourjopoulos, Nikos Fakotakis:
Automatic speech recognition performance in different room acoustic environments with and without dereverberation preprocessing. 380-395
Volume 27, Number 2, February 2013
- Sankaranarayanan Ananthakrishnan, Rohit Prasad, David Stallard, Prem Natarajan:
Batch-mode semi-supervised active learning for statistical machine translation. 397-406 - Xiaodong Cui, Mohamed Afify, Yuqing Gao, Bowen Zhou:
Stereo hidden Markov modeling for noise robust speech recognition. 407-419 - John Dines, Hui Liang, Lakshmi Babu Saheer, Matthew Gibson, William Byrne, Keiichiro Oura, Keiichi Tokuda, Junichi Yamagishi, Simon King, Mirjam Wester, Teemu Hirsimäki, Reima Karhila, Mikko Kurimo:
Personalising speech-to-speech translation: Unsupervised cross-lingual speaker adaptation for HMM-based speech synthesis. 420-437 - Emil Ettelaie, Panayiotis G. Georgiou, Shrikanth S. Narayanan:
Unsupervised data processing for classifier-based speech translator. 438-454 - Matthias Paulik, Alex Waibel:
Training speech translation from audio recordings of interpreter-mediated communication. 455-474 - Rohit Prasad, Prem Natarajan, David Stallard, Shirin Saleem, Shankar Ananthakrishnan, Stavros Tsakalidis, Chia-Lin Kao, Fred Choi, Ralf Meermeier, Mark Rawls, Jacob Devlin, Kriste Krstovski, Aaron Challenner:
BBN TransTalk: Robust multilingual two-way speech-to-speech translation for mobile platforms. 475-491 - Vivek Kumar Rangarajan Sridhar, Srinivas Bangalore, Shrikanth S. Narayanan:
Enriching machine-mediated speech-to-speech translation using contextual information. 492-508 - Sakriani Sakti, Michael Paul, Andrew M. Finch, Shinsuke Sakai, Thang Tat Vu, Noriyuki Kimura, Chiori Hori, Eiichiro Sumita, Satoshi Nakamura, Jun Park, Chai Wutiwiwatchai, Bo Xu, Hammam Riza, Karunesh Arora, Chi Mai Luong, Haizhou Li:
A-STAR: Toward translating Asian spoken languages. 509-527 - Gregory A. Sanders, Brian A. Weiss, Craig Schlenoff, Michelle Potts Steves, Sherri L. Condon:
Evaluation methodology and metrics employed to assess the TRANSTAC two-way, speech-to-speech translation systems. 528-553 - JongHo Shin, Panayiotis G. Georgiou, Shrikanth S. Narayanan:
Enabling effective design of multimodal interfaces for speech-to-speech translation system: An empirical study of longitudinal user behaviors over time and user strategies for coping with errors. 554-571 - Andreas Tsiartas, Prasanta Kumar Ghosh, Panayiotis G. Georgiou, Shrikanth S. Narayanan:
High-quality bilingual subtitle document alignments with application to spontaneous speech translation. 572-591 - Bowen Zhou, Xiaodong Cui, Songfang Huang, Martin Cmejrek, Wei Zhang, Jian Xue, Jia Cui, Bing Xiang, Gregg Daggett, Upendra V. Chaudhari, Sameer Maskey, Etienne Marcheret:
The IBM speech-to-speech translation system for smartphone: Improvements for resource-constrained tasks. 592-618
Volume 27, Number 3, May 2013
- Jon Barker, Emmanuel Vincent:
Special issue on speech separation and recognition in multisource environments. 619-620
- Jon Barker, Emmanuel Vincent, Ning Ma, Heidi Christensen, Phil D. Green:
The PASCAL CHiME speech separation and recognition challenge. 621-633 - Tania Habib, Harald Romsdorfer:
Auditory inspired methods for localization of multiple concurrent speakers. 634-659 - Alessio Brutti, Francesco Nesta:
Tracking of multidimensional TDOA for multiple sources with distributed microphone pairs. 660-682 - Pasi Pertilä:
Online blind speech separation using multiple acoustic speaker tracking and time-frequency masking. 683-702 - Francesco Nesta, Marco Matassoni:
Blind source extraction for robust speech recognition in multisource noisy environments. 703-725 - Klaus Reindl, Yuanhang Zheng, Andreas Schwarz, Stefan Meier, Roland Maas, Armin Sehr, Walter Kellermann:
A stereophonic acoustic signal extraction scheme for noisy and reverberant environments. 726-745 - Emad M. Grais, Hakan Erdogan:
Regularized nonnegative matrix factorization using Gaussian mixture priors for supervised single channel source separation. 746-762 - Antti Hurmalainen, Jort F. Gemmeke, Tuomas Virtanen:
Modelling non-stationary noise with spectral factorisation in automatic speech recognition. 763-779 - Martin Wöllmer, Felix Weninger, Jürgen T. Geiger, Björn W. Schuller, Gerhard Rigoll:
Noise robust ASR in reverberated multisource environments applying convolutive NMF and Long Short-Term Memory. 780-797 - Sami Keronen, Heikki Kallasjoki, Ulpu Remes, Guy J. Brown, Jort F. Gemmeke, Kalle J. Palomäki:
Mask estimation and imputation methods for missing data speech recognition in a multisource reverberant environment. 798-819 - Ning Ma, Jon Barker, Heidi Christensen, Phil D. Green:
A hearing-inspired approach for distant-microphone speech recognition in the presence of multiple sources. 820-836 - Ramón Fernandez Astudillo, Dorothea Kolossa, Alberto Abad, Steffen Zeiler, Rahim Saeidi, Pejman Mowlaee, João Paulo da Silva Neto, Rainer Martin:
Integration of beamforming and uncertainty-of-observation techniques for robust ASR in multi-source environments. 837-850 - Marc Delcroix, Keisuke Kinoshita, Tomohiro Nakatani, Shoko Araki, Atsunori Ogawa, Takaaki Hori, Shinji Watanabe, Masakiyo Fujimoto, Takuya Yoshioka, Takanobu Oba, Yotaro Kubo, Mehrez Souden, Seong-Jun Hahm, Atsushi Nakamura:
Speech recognition in living rooms: Integrated speech enhancement and recognition system based on spatial, spectral and temporal modeling of sounds. 851-873 - Alexey Ozerov, Mathieu Lagrange, Emmanuel Vincent:
Uncertainty-based learning of acoustic models from noisy data. 874-894
Volume 27, Number 4, June 2013
- Ilana Volfin, Israel Cohen:
Dominant speaker identification for multipoint videoconferencing. 895-910 - Xunying Liu, Mark J. F. Gales, Philip C. Woodland:
Language model cross adaptation for LVCSR system combination. 928-942 - Fuji Ren, Xin Kang:
Employing hierarchical Bayesian networks in simple and complex emotion topic analysis. 943-968 - Meng Sun, Hugo Van hamme:
Joint training of non-negative Tucker decomposition and discrete density hidden Markov models. 969-988 - Brad H. Story:
Phrase-level speech simulation with an airway modulation model of speech production. 989-1010 - David Wang, Robert Vogt, Sridha Sridharan:
Eigenvoice modelling for cross likelihood ratio based speaker clustering: A Bayesian approach. 1011-1027 - John Kane, Thomas Drugman, Christer Gobl:
Improved automatic detection of creak. 1028-1047
Volume 27, Number 5, August 2013
- Janne V. Kujala:
A probabilistic approach to pronunciation by analogy. 1049-1067 - Finnian Kelly, Andrzej Drygajlo, Naomi Harte:
Speaker verification in score-ageing-quality classification space. 1068-1084 - Ali Basirat, Heshaam Faili:
Bridge the gap between statistical and hand-crafted grammars. 1085-1104 - V. Ramu Reddy, K. Sreenivasa Rao:
Two-stage intonation modeling using feedforward neural networks for syllable based text-to-speech synthesis. 1105-1126 - Thomas Pellegrini, Rui Correia, Isabel Trancoso, Jorge Baptista, Nuno J. Mamede, Maxine Eskénazi:
ASR-based exercises for listening comprehension practice in European Portuguese. 1127-1142
Volume 27, Number 6, September 2013
- Kathleen F. McCoy, John L. Arnott, Leo Ferres, Melanie Fried-Oken, Brian Roark:
Speech and Language processing as assistive technologies. 1143-1146 - Harsh Vardhan Sharma, Mark Hasegawa-Johnson:
Acoustic model adaptation using in-domain background models for dysarthric speech recognition. 1147-1162 - Frank Rudzicz:
Adjusting dysarthric speech signals to be more intelligible. 1163-1177 - Sarah M. Creer, Stuart P. Cunningham, Phil D. Green, Junichi Yamagishi:
Building personalised synthetic voices for individuals with severe speech impairment. 1178-1193 - John L. Arnott, Norman Alm:
Towards the improvement of Augmentative and Alternative Communication through the modelling of conversation. 1194-1211 - Brian Roark, Russell Beckley, Chris Gibbons, Melanie Fried-Oken:
Huffman scanning: Using language models within fixed-grid keyboard emulation. 1212-1234 - Alberto Abad, Anna Pompili, Ângela Costa, Isabel Trancoso, José G. Fonseca, Gabriela Leal, Luisa Farrajota, Isabel P. Martins:
Automatic word naming recognition for an on-line aphasia treatment system. 1235-1248
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.