default search action
INTERSPEECH 2013: Lyon, France
- Frédéric Bimbot, Christophe Cerisara, Cécile Fougeron, Guillaume Gravier, Lori Lamel, François Pellegrino, Pascal Perrier:
14th Annual Conference of the International Speech Communication Association, INTERSPEECH 2013, Lyon, France, August 25-29, 2013. ISCA 2013
Systems for Search/Retrieval of Speech Documents
- Xavier Anguera:
Information retrieval-based dynamic time warping. 1-5 - Dogan Can, Shrikanth S. Narayanan:
On the computation of document frequency statistics from spoken corpora using factor automata. 6-10 - Kouichi Katsurada, Seiichi Miura, Kheang Seng, Yurie Iribe, Tsuneo Nitta:
Acceleration of spoken term detection using a suffix array by assigning optimal threshold values to sub-keywords. 11-14 - Arindam Mandal, Julien van Hout, Yik-Cheung Tam, Vikramjit Mitra, Yun Lei, Jing Zheng, Dimitra Vergyri, Luciana Ferrer, Martin Graciarena, Andreas Kathol, Horacio Franco:
Strategies for high accuracy keyword detection in noisy channels. 15-19 - Alberto Abad, Luis Javier Rodríguez-Fuentes, Mikel Peñagarikano, Amparo Varona, Germán Bordel:
On the calibration and fusion of heterogeneous spoken term detection systems. 20-24 - Shiro Narumi, Kazuma Konno, Takuya Nakano, Yoshiaki Itoh, Kazunori Kojima, Masaaki Ishigame, Kazuyo Tanaka, Shi-wook Lee:
Intensive acoustic models constructed by integrating low-occurrence models for spoken term detection. 25-28
Speech Analysis I-IV
- John Kane, Irena Yanushevskaya, John Dalton, Christer Gobl, Ailbhe Ní Chasaide:
Using phonetic feature extraction to determine optimal speech regions for maximising the effectiveness of glottal source analysis. 29-33 - Hideki Kawahara, Masanori Morise, Tomoki Toda, Ryuichi Nisimura, Toshio Irino:
Beyond bandlimited sampling of speech spectral envelope imposed by the harmonic structure of voiced sounds. 34-38 - JeeSok Lee, Frank K. Soong, Hong-Goo Kang:
A source-filter based adaptive harmonic model and its application to speech prosody modification. 39-43 - K. Ramesh, S. R. Mahadeva Prasanna, D. Govind:
Detection of glottal opening instants using Hilbert envelope. 44-48 - Dhananjaya N. Gowda, Jouni Pohjalainen, Mikko Kurimo, Paavo Alku:
Robust formant detection using group delay function and stabilized weighted linear prediction. 49-53 - Thomas Hézard, Thomas Hélie, Boris Doval:
A source-filter separation algorithm for voiced sounds based on an exact anticausal/causal pole decomposition for the class of periodic signals. 54-58
Language and Dialect Recognition
- Weiwei Liu, Wei-Qiang Zhang, Zhiyi Li, Jia Liu:
Parallel absolute-relative feature based phonotactic language recognition. 59-63 - Mireia Díez, Amparo Varona, Mikel Peñagarikano, Luis Javier Rodríguez-Fuentes, Germán Bordel:
Dimensionality reduction of phone log-likelihood ratio features for spoken language recognition. 64-68 - Jeff Z. Ma, Bing Zhang, Spyros Matsoukas, Sri Harish Reddy Mallidi, Feipeng Li, Hynek Hermansky:
Improvements in language identification on the RATS noisy speech corpus. 69-73 - Mehdi Soufifar, Lukás Burget, Oldrich Plchot, Sandro Cumani, Jan Cernocký:
Regularized subspace n-gram model for phonotactic ivector extraction. 74-78 - Hamid Behravan, Ville Hautamäki, Tomi Kinnunen:
Foreign accent detection from spoken Finnish using i-vectors. 79-83 - Mitchell McLaren, Aaron Lawson, Yun Lei, Nicolas Scheffer:
Adaptive Gaussian backend for robust language identification. 84-88
ASR - Neural Networks
- Matthias Paulik:
Lattice-based training of bottleneck feature extraction neural networks. 89-93 - Jonas Gehring, Wonkyum Lee, Kevin Kilgour, Ian R. Lane, Yajie Miao, Alex Waibel:
Modular combination of deep neural networks for acoustic modeling. 94-98 - Shuo-Yiin Chang, Nelson Morgan:
Informative spectro-temporal bottleneck features for noise-robust speech recognition. 99-103 - Zhi-Jie Yan, Qiang Huo, Jian Xu:
A scalable approach to using DNN-derived features in GMM-HMM based acoustic modeling for LVCSR. 104-108 - Shakti P. Rath, Daniel Povey, Karel Veselý, Jan Cernocký:
Improved feature processing for deep neural networks. 109-113 - Oriol Vinyals, Nelson Morgan:
Deep vs. wide: depth on a budget for robust speech recognition. 114-118
Speech Acoustics
- Angelika Braun:
An early case of "VOT". 119-122 - Robert Allen Fox, Ewa Jacewicz, Jessica Hart:
Pitch pattern variations in three regional varieties of American English. 123-127 - Jean-Sylvain Liénard, Claude Barras:
Fine-grain voice strength estimation from vowel spectral cues. 128-132 - Elizabeth Godoy, Catherine Mayo, Yannis Stylianou:
Linking loudness increases in normal and lombard speech to decreasing vowel formant separation. 133-137 - Kunitoshi Motoki:
Three-dimensional rectangular vocal-tract model with asymmetric wall impedances. 138-142 - Manu Airaksinen, Brad H. Story, Paavo Alku:
Quasi closed phase analysis for glottal inverse filtering. 143-147
Paralinguistic Challenge (Special Session)
- Björn W. Schuller, Stefan Steidl, Anton Batliner, Alessandro Vinciarelli, Klaus R. Scherer, Fabien Ringeval, Mohamed Chetouani, Felix Weninger, Florian Eyben, Erik Marchi, Marcello Mortillaro, Hugues Salamin, Anna Polychroniou, Fabio Valente, Samuel Kim:
The INTERSPEECH 2013 computational paralinguistics challenge: social signals, conflict, emotion, autism. 148-152 - Artur Janicki:
Non-linguistic vocalisation recognition based on hybrid GMM-SVM approach. 153-157 - Jieun Oh, Eunjoon Cho, Malcolm Slaney:
Characteristic contours of syllabic-level units in laughter. 158-162 - Teun F. Krikke, Khiet P. Truong:
Detection of nonverbal vocalizations using Gaussian mixture models: looking for fillers and laughter in conversational speech. 163-167 - Johannes Wagner, Florian Lingenfelser, Elisabeth André:
Using phonetic patterns for detecting social cues in natural conversations. 168-172 - Rahul Gupta, Kartik Audhkhasi, Sungbok Lee, Shrikanth S. Narayanan:
Paralinguistic event detection from speech using probabilistic time-series smoothing and masking. 173-177 - Gouzhen An, David Guy Brizan, Andrew Rosenberg:
Detecting laughter and filled pauses using syllable-based features. 178-181 - Daniel Bone, Theodora Chaspari, Kartik Audhkhasi, James Gibson, Andreas Tsiartas, Maarten Van Segbroeck, Ming Li, Sungbok Lee, Shrikanth S. Narayanan:
Classifying language-related developmental disorders from speech cues: the promise and the potential confounds. 182-186 - Katrin Kirchhoff, Yuzong Liu, Jeff A. Bilmes:
Classification of developmental disorders from speech signals using submodular feature selection. 187-190 - Meysam Asgari, Alireza Bayestehtashk, Izhak Shafran:
Robust and accurate features for detecting and diagnosing autism spectrum disorders. 191-194 - David Martínez González, Dayana Ribas, Eduardo Lleida, Alfonso Ortega, Antonio Miguel:
Suprasegmental information modelling for autism disorder spectrum and specific language impairment classification. 195-199 - Félix Grèzes, Justin Richards, Andrew Rosenberg:
Let me finish: automatic conflict detection using speaker overlap. 200-204 - Vidhyasaharan Sethu, Julien Epps, Eliathamby Ambikairajah, Haizhou Li:
GMM based speaker variability compensated system for interspeech 2013 compare emotion challenge. 205-209 - Okko Räsänen, Jouni Pohjalainen:
Random subset feature selection in automatic recognition of developmental disorders, affective states, and level of conflict from speech. 210-214 - Hung-yi Lee, Ting-Yao Hu, How Jing, Yun-Fan Chang, Yu Tsao, Yu-Cheng Kao, Tsang-Long Pao:
Ensemble of machine learning and acoustic segment model techniques for speech emotion and autism spectrum disorders recognition. 215-219 - Gábor Gosztolya, Róbert Busa-Fekete, László Tóth:
Detecting autism, emotions and social signals using adaboost. 220-224
Perception of Prosody
- Oliver Niebuhr:
Resistance is futile - the intonation between continuation rise and calling contour in German. 225-229 - Hansjörg Mixdorff, Oliver Niebuhr:
The influence of F0 contour continuity on prominence perception. 230-234 - Caroline L. Smith, Paul Edmunds:
Native English listeners' perceptions of prosody in L1 and L2 reading. 235-238 - Chiharu Tsurutani, Dean Luo:
Naturalness judgement of L2 Mandarin Chinese - does timing matter? 239-242 - Daniel Aalto, Juraj Simko, Martti Vainio:
Language background affects the strength of the pitch bias in a duration discrimination task. 243-247 - Margaret Zellers:
Pitch and lengthening as cues to turn transition in Swedish. 248-252 - Maria Paola Bissiri, Margaret Zellers:
Perception of glottalization in varying pitch contexts across languages. 253-257 - Michael Walsh, Katrin Schweitzer, Nadja Schauffler:
Exemplar-based pitch accent categorisation using the generalized context model. 258-262 - Bettina Braun, Yuki Asano:
Double contrast is signalled by prenuclear and nuclear accent types alone, not by f0-plateaux. 263-266 - Susana Correia, Sónia Frota, Joseph Butler, Marina Vigário:
Word stress perception in European Portuguese. 267-271 - Denis Arnold, Petra Wagner, R. Harald Baayen:
Using generalized additive models and random forests to model prosodic prominence in German. 272-276 - Hartmut R. Pfitzinger, Hansjörg Mixdorff:
Perceiving speech rate differences between natural and time-scale modified utterances. 277-281
Prosody, Phonetics of Language Varieties
- Plínio A. Barbosa, Anders Eriksson, Joel Åkesson:
On the robustness of some acoustic parameters for signalling word stress across styles in Brazilian Portuguese. 282-286 - Shao-Ren Lyu, Ho-hsien Pan:
Reexamine the sandhi rules and the merging tones in hakka language. 287-290 - Marija Tabain, Richard Beare, Andrew Butcher:
A preliminary spectral analysis of palatal and velar stop bursts in pitjantjatjara. 291-295 - Shakuntala Mahanta, A. I. Twaha:
Presentational focus realisation in nalbaria variety of assamese. 296-299 - Marisa Cruz, Sónia Frota:
On the relation between intonational phrasing and pitch accent distribution. evidence from European Portuguese varieties. 300-304 - Rena Nemoto, Martine Adda-Decker:
How are word-final schwas different in the north and south of france? 305-309 - Simone Ashby, Sílvia Barbosa, Catarina Silva, Paulino Fumo, José Pedro Ferreira:
Modeling postcolonial language varieties: challenges and lessons learned from mozambican Portuguese. 310-314 - Heete Sahkai, Mari-Liis Kalvik, Meelis Mihkla:
Prosody of contrastive focus in estonian. 315-319 - Thomas Kisler, Uwe D. Reichel:
Exploring the connection of acoustic and distinctive features. 320-324 - Conceição Cunha, Jonathan Harrington, Phil Hoole:
A physiological analysis of the tense/lax vowel contrast in two varieties of German. 325-329 - Einar Meister, Lya Meister:
Production of estonian quantity contrasts by native speakers of Finnish. 330-334 - Yohann Meynadier, Yulia Gaydina:
Aerodynamic and durational cues of phonological voicing in whisper. 335-339 - Uwe D. Reichel:
Information theoretic syllable structure and its relation to the c-center effect. 340-344 - Bistra Andreeva, William J. Barry, Jacques C. Koreman:
The bulgarian stressed and unstressed vowel system. a corpus study. 345-348
Speech Synthesis I. II
- Santitham Prom-on, Peter Birkholz, Yi Xu:
Training an articulatory synthesizer with continuous acoustic data. 349-353 - Géza Kiss, Jan P. H. van Santen:
Estimating speaker-specific intonation patterns using the linear alignment model. 354-358 - June Sig Sung, Doo Hwa Hong, Hyun Woo Koo, Nam Soo Kim:
Factored maximum likelihood kernelized regression for HMM-based singing voice synthesis. 359-363 - Shinnosuke Takamichi, Tomoki Toda, Yoshinori Shiga, Sakriani Sakti, Graham Neubig, Satoshi Nakamura:
Improvements to HMM-based speech synthesis based on parameter generation with rich context models. 364-368 - Toru Nakashika, Ryoichi Takashima, Tetsuya Takiguchi, Yasuo Ariki:
Voice conversion in high-order eigen space using deep belief nets. 369-372 - Hanna Silén, Jani Nurminen, Elina Helander, Moncef Gabbouj:
Voice conversion for non-parallel datasets using dynamic kernel partial least squares regression. 373-377 - Takashi Nose, Misa Kanemoto, Tomoki Koriyama, Takao Kobayashi:
A style control technique for singing voice synthesis based on multiple-regression HSMM. 378-382 - Florian Hinterleitner, Christoph Norrenbrock, Sebastian Möller, Ulrich Heute:
Predicting the quality of text-to-speech systems from a large-scale feature set. 383-387 - Jani Nurminen, Hanna Silén, Moncef Gabbouj:
Speaker-specific retraining for enhanced compression of unit selection text-to-speech databases. 388-391 - Mark A. Huckvale, Julian Leff, Geoff Williams:
Avatar therapy: an audio-visual dialogue system for treating auditory hallucinations. 392-396 - Prasanna Kumar Muthukumar, Alan W. Black, H. Timothy Bunnell:
Optimizations and fitting procedures for the liljencrants-fant model for statistical parametric speech synthesis. 397-401 - Dirk Hovy, Gopala Krishna Anumanchipalli, Alok Parlikar, Caroline Vaughn, Adam C. Lammert, Eduard H. Hovy, Alan W. Black:
Analysis and modeling of "focus" in context. 402-406
Perception, Dialectal Differences
- Thi Anh Xuan Tran, Viet Son Nguyen, Eric Castelli, René Carré:
Production and perception of pseudo-V1CV2 outside the vowel triangle: speech illusion effects. 407-411 - Maria Candea, Martine Adda-Decker, Lori Lamel:
Recent evolution of non-standard consonantal variants in French broadcast news. 412-416 - Frank Zimmerer, Rei Yasuda, Henning Reetz:
Architekt or archtekt? perception of devoiced vowels produced by Japanese speakers of German. 417-420 - Andrew R. Plummer, Lucie Ménard, Benjamin Munson, Mary E. Beckman:
Comparing vowel category response surfaces over age-varying maximal vowel spaces within and across language communities. 421-425 - Molly Babel, Grant McGuire:
Perceived vocal attractiveness across dialects is similar but not uniform. 426-430 - Hongyan Wang, Vincent J. van Heuven:
Mutual intelligibility of American, Chinese and Dutch-accented speakers of English tested by SUS and SPIN sentences. 431-435
Speech Enhancement - Single Channel
- Xugang Lu, Yu Tsao, Shigeki Matsuda, Chiori Hori:
Speech enhancement based on deep denoising autoencoder. 436-440 - Hiroshi Saruwatari, Suzumi Kanehara, Ryoichi Miyazaki, Kiyohiro Shikano, Kazunobu Kondo:
Musical noise analysis for Bayesian minimum mean-square error speech amplitude estimators based on higher-order statistics. 441-445 - Nikolay Lyubimov, Mikhail Kotov:
Non-negative matrix factorization with linear constraints for single-channel speech enhancement. 446-450 - Hung-Wei Tseng, Srikanth Vishnubhotla, Mingyi Hong, Xiangfeng Wang, Jinjun Xiao, Zhi-Quan Luo, Tao Zhang:
A single channel speech enhancement approach by combining statistical criterion and multi-frame sparse dictionary learning. 451-455 - Majid Mirbagheri, Yanbo Xu, Sahar Akram, Shihab A. Shamma:
Speech enhancement using convolutive nonnegative matrix factorization with cosparsity regularization. 456-459 - Matthew C. McCallum, Bernard J. Guillemin:
Joint stochastic-deterministic wiener filtering with recursive Bayesian estimation of deterministic speech. 460-464
Dialog Modeling
- Juha Knuuttila, Okko Räsänen, Unto K. Laine:
Automatic self-supervised learning of associations between speech and text. 465-469 - Lucie Daubigney, Matthieu Geist, Olivier Pietquin:
Particle swarm optimisation of spoken dialogue system strategies. 470-474 - Pierre Lison:
Model-based Bayesian reinforcement learning for dialogue management. 475-479 - Fabrizio Ghigi, M. Inés Torres, Raquel Justo, José-Miguel Benedí:
Evaluating spoken dialogue models under the interactive pattern recognition framework. 480-484 - Yun-Nung Chen, Florian Metze:
Multi-layer mutually reinforced random walk with hidden parameters for improved multi-party meeting summarization. 485-489 - Pei-hao Su, Yow-Bang Wang, Tsung-Hsien Wen, Tien-han Yu, Lin-Shan Lee:
A recursive dialogue game framework with optimal Policy offering personalized computer-assisted language learning. 490-494
ASR - Lexical, Prosodic and Cross/Multi-Lingual
- Stefan Hahn, Patrick Lehnen, Simon Wiesler, Ralf Schlüter, Hermann Ney:
Improving LVCSR with hidden conditional random fields for grapheme-to-phoneme conversion. 495-499 - Van Hai Do, Xiong Xiao, Engsiong Chng, Haizhou Li:
Context-dependent phone mapping for LVCSR of under-resourced languages. 500-504 - Ramya Rasipuram, Mathew Magimai-Doss:
Improving grapheme-based ASR by probabilistic lexical modeling approach. 505-509 - Petr Motlícek, David Imseng, Philip N. Garner:
Crosslingual tandem-SGMM: exploiting out-of-language data for acoustic model and feature level adaptation. 510-514 - Ngoc Thang Vu, Tanja Schultz:
Multilingual multilayer perceptron for rapid language adaptation between and across language families. 515-519 - Andrew Rosenberg:
Modeling prosodic sequences with k-means and dirichlet process GMMs. 520-524
Phonetic Convergence
- Antje Schweitzer, Natalie Lewandowski:
Convergence of articulation rate in spontaneous speech. 525-529 - Jennifer S. Pardo:
Phonetic convergence in shadowed speech: a comparison of perceptual and acoustic measures. 530-534 - Marcin Wlodarczak, Juraj Simko, Petra Wagner:
Pitch and duration as a basis for entrainment of overlapped speech onsets. 535-538 - Francesca Bonin, Céline De Looze, Sucheta Ghosh, Emer Gilmartin, Carl Vogel, Anna Polychroniou, Hugues Salamin, Alessandro Vinciarelli, Nick Campbell:
Investigating fine temporal dynamics of prosodic and lexical accommodation. 539-543 - Jeesun Kim, Ruben Demirdjian, Chris Davis:
Spontaneous and explicit speech imitation. 544-547 - Václav Jonás Podlipský, Sárka Simácková, Katerina Chládková:
Imitation interacts with one's second-language phonology but it does not operate cross-linguistically. 548-552
Speech Production, Acquisition and Development I, II
- Po-jen Hsieh:
Prosodic markings of semantic predictability in taiwan Mandarin. 553-557 - Rüdiger Hoffmann, Dieter Mehnert, Rolf Dietzel:
How did it work? historic phonetic devices explained by coeval photographs. 558-562 - Lea S. Kohtz, Oliver Niebuhr:
Eliciting speech with sentence lists - a critical evaluation with special emphasis on segmental anchoring. 563-567 - Yuguang Wang, Jianwu Dang, Xi Chen, Jianguo Wei, Hongcui Wang, Kiyoshi Honda:
An MRI-based acoustic study of Mandarin vowels. 568-571 - Daniel Hirst:
Melody metrics for prosodic typology: comparing English, French and Chinese. 572-576 - Michael I. Proctor, Louis Goldstein, Adam C. Lammert, Dani Byrd, Asterios Toutios, Shrikanth S. Narayanan:
Velic coordination in French nasals: a real-time magnetic resonance imaging study. 577-581 - Mark A. Huckvale, Amrita Sharma:
Learning to imitate adult speech with the KLAIR virtual infant. 582-586 - Jorge C. Lucero, Jean Schoentgen, Mara Behlau:
Physics-based synthesis of disordered voices. 587-591 - Sonia D'Apolito, Barbara Gili Fivela:
Place assimilation and articulatory strategies: the case of sibilant sequences in French as L1 and L2. 592-596 - Barbara Samlowski, Petra Wagner, Bernd Möbius:
Effects of lexical class and lemma frequency on German homographs. 597-601 - Leonardo Lancia, Heriberto Avelino, Daniel Voigt:
Measuring laryngealization in running speech: interaction with contrastive tones in yalálag zapotec. 602-606 - Erin Rusaw:
A neural oscillator model of speech timing and rhythm. 607-611 - Nicole Wong, Maojing Fu, Zhi-Pei Liang, Ryan Shosted, Bradley P. Sutton:
Observations of perseverative coarticulation in lateral approximants using MRI. 612-616
General Topics in ASR
- Vishwa Gupta, Gilles Boulianne:
Comparing computation in Gaussian mixture and neural network based large-vocabulary speech recognition. 617-621 - Daniel Stein, Jochen Schwenninger, Michael Stadtschnitzer:
Simultaneous perturbation stochastic approximation for automatic speech recognition. 622-626 - David Sheffield, Michael J. Anderson, Yunsup Lee, Kurt Keutzer:
Hardware/software codesign for mobile speech recognition. 627-631 - Yangyang Shi, Martha A. Larson, Pascal Wiggers, Catholijn M. Jonker:
Exploiting the succeeding words in recurrent neural network language models. 632-636 - Amir Hossein Harati Nejad Torbati, Joseph Picone, Marc Sobel:
Speech acoustic unit segmentation using hierarchical dirichlet processes. 637-641 - Munir Georges, Stephan Kanthak, Dietrich Klakow:
Transducer-based speech recognition with dynamic language models. 642-646 - Yotaro Kubo, Takaaki Hori, Atsushi Nakamura:
A method for structure estimation of weighted finite-state transducers and its application to grapheme-to-phoneme conversion. 647-651 - Denis Jouvet, Dominique Fohr:
Combining forward-based and backward-based decoders for improved speech recognition performance. 652-656 - Olivier Siohan, Michiel Bacchiani:
ivector-based acoustic data selection. 657-661 - Xin Lei, Andrew W. Senior, Alexander Gruenstein, Jeffrey Sorensen:
Accurate and compact large vocabulary speech recognition on mobile devices. 662-665 - Cyril Allauzen, Michael Riley:
Pre-initialized composition for large-vocabulary speech recognition. 666-670 - Evelyn Kurniawati, Sapna George:
Speaker dependent activation keyword detector based on GMM-UBM. 671-674 - Hasim Sak, Yun-Hsuan Sung, Françoise Beaufays, Cyril Allauzen:
Written-domain language modeling for automatic speech recognition. 675-679
Voice Activity Detection and Speech Segmentation
- Maarten Versteegh, Louis ten Bosch:
Detecting words in speech using linear separability in a bag-of-events vector space. 680-684 - Matt Burlick, Dimitrios Dimitriadis, Eric Zavesky:
On the improvement of multimodal voice activity detection. 685-689 - Jürgen T. Geiger, Florian Eyben, Nicholas W. D. Evans, Björn W. Schuller, Gerhard Rigoll:
Using linguistic information to detect overlapping speech. 690-694 - Jiaxing Ye, Takumi Kobayashi, Masahiro Murakawa, Tetsuya Higuchi:
Incremental acoustic subspace learning for voice activity detection using harmonicity-based features. 695-699 - Hoon Chung, Sung Joo Lee, Yunkeun Lee:
Endpoint detection using weighted finite state transducer. 700-703 - Maarten Van Segbroeck, Andreas Tsiartas, Shrikanth S. Narayanan:
A robust frontend for VAD: exploiting contextual, discriminative and spectral cues of human voice. 704-708 - Martin Graciarena, Abeer Alwan, Dan Ellis, Horacio Franco, Luciana Ferrer, John H. L. Hansen, Adam Janin, Byung Suk Lee, Yun Lei, Vikramjit Mitra, Nelson Morgan, Seyed Omid Sadjadi, T. J. Tsai, Nicolas Scheffer, Lee Ngee Tan, Benjamin Williams:
All for one: feature combination for highly channel-degraded speech activity detection. 709-713 - Maxime Le Coz, Julien Pinquier, Régine André-Obrecht:
Superposed speech localisation using frequency tracking. 714-717 - Andreas Tsiartas, Theodora Chaspari, Nassos Katsamanis, Prasanta Kumar Ghosh, Ming Li, Maarten Van Segbroeck, Alexandros Potamianos, Shrikanth S. Narayanan:
Multi-band long-term signal variability features for robust voice activity detection. 718-722 - Narimene Lezzoum, Ghyslain Gagnon, Jérémie Voix:
A low-complexity voice activity detector for smart hearing protection of hyperacusic persons. 723-727 - Neville Ryant, Mark Liberman, Jiahong Yuan:
Speech activity detection on youtube using deep neural networks. 728-731 - François G. Germain, Dennis L. Sun, Gautham J. Mysore:
Speaker and noise independent voice activity detection. 732-736 - T. J. Tsai, Adam Janin:
Confidence-based scoring: a useful diagnostic tool for detection tasks. 737-741 - Yasuaki Kanai, Shota Morita, Masashi Unoki:
Concurrent processing of voice activity detection and noise reduction using empirical mode decomposition and modulation spectrum analysis. 742-746
Show and Tell Sessions 1-3
- Samer Al Moubayed, Jonas Beskow, Gabriel Skantze:
The furhat social companion talking head. 747-749 - Rodolphe Gelin, Gabriele Barbieri:
Audition: the most important sense for humanoid robots? 750-751 - Thomas Hueber:
Ultraspeech-player: intuitive visualization of ultrasound articulatory data for speech therapy and pronunciation training. 752-753 - Jieun Oh, Ge Wang:
Laughter modulation: from speech to speech-laugh. 754-755 - Daniel M. Bikel, Keith B. Hall:
Refr: an open-source reranker framework. 756-758 - Alessandro Sosi, Fabio Brugnara, Luca Cristoforetti, Marco Matassoni, Mirco Ravanelli, Maurizio Omologo:
Embedding speech recognition to control lights. 759-760 - Geoffrey S. Meltzner, James T. Heaton, Yunbin Deng:
The MUTE silent speech recognition system. 761-763 - James M. Scobbie, Alice Turk, Christian Geng, Simon King, Robin J. Lickley, Korin Richmond:
The edinburgh speech production facility doubletalk corpus. 764-766 - Dmitry Sityaev, Jonathan Hotz, Vadim Snitkovsky:
Lexee: a cloud-based platform for building and deploying voice-enabled mobile applications. 767-769 - Slim Ouni:
Visualizing articulatory data with VisArtico. 770-772 - Mariette Soury, Clément Gossart, Martine Adda-Decker, Laurence Devillers:
A tool to elicit and collect multicultural and multimodal laughter. 773-774 - Robert Schleicher, Tilo Westermann, Jinjin Li, Moritz Lawitschka, Benjamin Mateev, Ralf Reichmuth, Sebastian Möller:
Design of a mobile app for interspeech conferences: towards an open tool for the spoken language community. 775-777
Discourse, Intonation, Prosody
- Anders Eriksson, Plínio A. Barbosa, Joel Åkesson:
The acoustics of word stress in Swedish: a function of stress level, speaking style and word accent. 778-782 - Amandine Michelas, Cristel Portes, Maud Champagne-Lavau:
Intonational contrasts encode speaker's certainty in neutral vs. incredulity declarative questions in French. 783-787 - Yuichi Ishimoto, Mika Enomoto, Hitoshi Iida:
Prosodic changes pre-announcing a syntactic completion point in Japanese utterance. 788-792 - Candide Simard:
Prosodic encoding of declarative, interrogative and imperative sentences in jaminjung, a language of australia. 793-797 - Anne Vullinghs, Martijn Goudbeek, Emiel Krahmer:
Crosslinguistic priming in interactive reference: evidence for conceptual alignment in speech production. 798-802 - Spyros Kousidis, David Schlangen, Stavros Skopeteas:
A cross-linguistic study on turn-taking and temporal alignment in verbal interaction. 803-807
Source Separation
- Emad M. Grais, Hakan Erdogan:
Discriminative nonnegative dictionary learning using cross-coherence penalties for single channel source separation. 808-812 - Han-Gyu Kim, Gil-Jin Jang, Jeong-Sik Park, Yung-Hwan Oh:
Monaural speech segregation based on pitch track correction using an ensemble kalman filter. 813-816 - Ngoc Thuy Tran, William G. Cowley, André Pollok:
Voice activity classification for automatic bi-speaker adaptive beamforming in speech separation. 817-821 - Keisuke Kinoshita, Mehrez Souden, Tomohiro Nakatani:
Blind source separation using spatially distributed microphones based on microphone-location dependent source activities. 822-826 - Tom Barker, Tuomas Virtanen:
Non-negative tensor factorisation of modulation spectrograms for monaural sound source separation. 827-831 - Mario Kaoru Watanabe, Pejman Mowlaee:
Iterative sinusoidal-based partial phase reconstruction in single-channel source separation. 832-836
Paralinguistic Information I, II
- Xiao Yao, Takatoshi Jitsuhiro, Chiyomi Miyajima, Norihide Kitaoka, Kazuya Takeda:
Classification of speech under stress by modeling the aerodynamics of the laryngeal ventricle. 837-841 - Rachel Rakov, Andrew Rosenberg:
"sure, i did the right thing": a system for sarcasm detection in speech. 842-846 - Stefan Scherer, Giota Stratou, Jonathan Gratch, Louis-Philippe Morency:
Investigating voice quality as a speaker-independent indicator of depression and PTSD. 847-851 - Thomas Pellegrini, Annika Hämäläinen, Philippe Boula de Mareüil, Michael Tjalve, Isabel Trancoso, Sara Candeias, Miguel Sales Dias, Daniela Braga:
A corpus-based study of elderly and young speakers of European Portuguese: acoustic correlates and their impact on speech recognition performance. 852-856 - Nicholas Cummins, Julien Epps, Vidhyasaharan Sethu, Michael Breakspear, Roland Goecke:
Modeling spectral variability for the classification of depressed speech. 857-861 - Verónica Pérez-Rosas, Rada Mihalcea:
Sentiment analysis of online spoken reviews. 862-866
ASR - Robustness Against Noise I-III
- Ahmed Hussen Abdelaziz, Steffen Zeiler, Dorothea Kolossa:
Using twin-HMM-based audio-visual speech enhancement as a front-end for robust audio-visual speech recognition. 867-871 - James Gibson, Maarten Van Segbroeck, Antonio Ortega, Panayiotis G. Georgiou, Shrikanth S. Narayanan:
Spectro-temporal directional derivative features for automatic speech recognition. 872-875 - Xiong Xiao, Engsiong Chng, Haizhou Li:
Attribute-based histogram equalization (HEQ) and its adaptation for robust speech recognition. 876-880 - Vikas Joshi, N. Vishnu Prasad, Srinivasan Umesh:
Modified cepstral mean normalization - transforming to utterance specific non-zero mean. 881-885 - Vikramjit Mitra, Horacio Franco, Martin Graciarena:
Damped oscillator cepstral coefficients for robust speech recognition. 886-890 - Md. Jahangir Alam, Patrick Kenny, Douglas D. O'Shaughnessy:
Regularized MVDR spectrum estimation-based robust feature extractors for speech recognition. 891-895
Neural Basis of Speech Perception
- Víctor Poblete, Néstor Becerra Yoma, Richard M. Stern:
Optimization of sigmoidal rate-level function based on acoustic features. 896-900 - Makiko Sadakata, Loukianos Spyrou, Mizuki Shingai, Kaoru Sekiyama:
Composing auditory ERPs: cross-linguistic comparison of auditory change complex for Japanese fricative consonants. 901-905 - Nathalie Bedoin, Jennifer Krzonowski, Emmanuel Ferragne:
How voicing, place and manner of articulation differently modulate event-related potentials associated with response inhibition. 906-910 - Ludovic Bellier, Michel Mazzuca, Hung Thai-Van, Anne Caclin, Rafael Laboissière:
Categorization of speech in early auditory evoked responses. 911-915 - Anna Dora Manca, Mirko Grimaldi:
Perception and production of Italian vowels: an ERP study. 916-920 - Ann-Kathrin Grohe, Bettina Braun:
Implicit learning leads to familiarity effects for intonation but not for voice. 921-924
Spoofing and Countermeasures for Automatic Speaker Verification (Special Session)
- Nicholas W. D. Evans, Tomi Kinnunen, Junichi Yamagishi:
Spoofing and countermeasures for automatic speaker verification. 925-929 - Rosa González Hautamäki, Tomi Kinnunen, Ville Hautamäki, Timo Leino, Anne-Maria Laukkanen:
I-vectors meet imitators: on vulnerability of speaker verification systems against voice mimicry. 930-934 - Marta Gomez-Barrero, Javier Gonzalez-Dominguez, Javier Galbally, Joaquin Gonzalez-Rodriguez:
Security evaluation of i-vector based speaker verification systems against hill-climbing attacks. 935-939 - Federico Alegre, Ravichander Vipperla, Asmaa Amehraye, Nicholas W. D. Evans:
A new speaker verification spoofing countermeasure based on local binary patterns. 940-944 - Zvi Kons, Hagai Aronowitz:
Voice transformation-based spoofing of text-dependent speaker verification systems. 945-949 - Zhizheng Wu, Anthony Larcher, Kong-Aik Lee, Engsiong Chng, Tomi Kinnunen, Haizhou Li:
Vulnerability evaluation of speaker verification under voice conversion spoofing: the effect of text constraints. 950-954
Speech Production, Acquisition and Development I, II
- Masako Fujimoto, Tatsuya Kitamura, Hiroaki Hatano, Ichiro Fujimoto:
Timing differences in articulation between voiced and voiceless stop consonants: an analysis of cine-MRI data. 955-958 - Adam C. Lammert, Vikram Ramanarayanan, Michael I. Proctor, Shrikanth S. Narayanan:
Vocal tract cross-distance estimation from real-time MRI using region-of-interest analysis. 959-962 - Apoorv Reddy Arrabothu, Nivedita Chennupati, B. Yegnanarayana:
Syllable nuclei detection using perceptually significant features. 963-967 - Fang-Ying Hsieh, Louis Goldstein, Dani Byrd, Shrikanth S. Narayanan:
Truncation of pharyngeal gesture in English diphthong [aɪ]. 968-972 - Zhaojun Yang, Vikram Ramanarayanan, Dani Byrd, Shrikanth S. Narayanan:
The effect of word frequency and lexical class on articulatory-acoustic coupling. 973-977 - Kimiko Yamakawa, Shigeaki Amano:
Discrimination between fricative and affricate in Japanese using time and spectral domain variables. 978-981 - Polina Drozdova, Catia Cucchiarini, Helmer Strik:
L2 syntax acquisition: the effect of oral and written computer assisted practice. 982-986 - Rosario Signorello, Didier Demolin:
The physiological use of the charismatic voice in Political speech. 987-991 - Ralph L. Rose:
Crosslinguistic corpus of hesitation phenomena: a corpus for investigating first and second language speech performance. 992-996 - Simon Preuß, Christiane Neuschaefer-Rube, Peter Birkholz:
Real-time control of a 2d animation model of the vocal tract using optopalatography. 997-1001 - Jessica Siddins, Jonathan Harrington, Felicitas Kleber, Ulrich Reubold:
The influence of accentuation and polysyllabicity on compensatory shortening in German. 1002-1006 - Hongwei Ding, Rüdiger Hoffmann:
An investigation of vowel epenthesis in Chinese learners' production of German consonants. 1007-1011 - Korin Richmond, Zhen-Hua Ling, Junichi Yamagishi, Benigno Uria:
On the evaluation of inversion mapping performance in the acoustic domain. 1012-1016
Speech Synthesis I. II
- Tatsuma Ishihara, Hirokazu Kameoka, Kota Yoshizato, Daisuke Saito, Shigeki Sagayama:
Probabilistic speech F0 contour model incorporating statistical vocabulary model of phrase-accent command sequence. 1017-1021 - Ian Vince McLoughlin, Jingjie Li, Yan Song:
Reconstruction of continuous voiced speech from whispers. 1022-1026 - Daniel R. van Niekerk, Etienne Barnard:
Generating fundamental frequency contours for speech synthesis in yorùbá. 1027-1031 - Elias Azarov, Maxim Vashkevich, Denis Likhachov, Alexander A. Petrovsky:
Real-time voice conversion using artificial neural networks with rectified linear units. 1032-1036 - Oraphan Krityakien, Keikichi Hirose, Nobuaki Minematsu:
Generation of fundamental frequency contours for Thai speech synthesis using tone nucleus model. 1037-1041 - Langzhou Chen, Norbert Braunschweiler:
Unsupervised speaker and expression factorization for multi-speaker expressive synthesis of ebooks. 1042-1046 - Hideharu Nakajima, Hideyuki Mizuno, Osamu Yoshioka, Satoshi Takahashi:
Which resemblance is useful to predict phrase boundary rise labels for Japanese expressive text-to-speech synthesis, numerically-expressed stylistic or distribution-based semantic? 1047-1051 - Jinfu Ni, Yoshinori Shiga, Chiori Hori, Yutaka Kidawara:
A targets-based superpositional model of fundamental frequency contours applied to HMM-based speech synthesis. 1052-1056 - Kazuhiro Kobayashi, Hironori Doi, Tomoki Toda, Tomoyasu Nakano, Masataka Goto, Graham Neubig, Sakriani Sakti, Satoshi Nakamura:
An investigation of acoustic features for singing voice conversion based on perceptual age. 1057-1061 - Bajibabu Bollepalli, Tuomo Raitio, Paavo Alku:
Effect of MPEG audio compression on HMM-based speech synthesis. 1062-1066 - Hironori Doi, Tomoki Toda, Tomoyasu Nakano, Masataka Goto, Satoshi Nakamura:
Evaluation of a singing voice conversion method based on many-to-many eigenvoice conversion. 1067-1071 - Tomoki Koriyama, Takashi Nose, Takao Kobayashi:
Statistical nonparametric speech synthesis using sparse Gaussian processes. 1072-1076 - Amir Mohammadi, Cenk Demiroglu:
Hybrid nearest-neighbor/cluster adaptive training for rapid speaker adaptation in statistical speech synthesis systems. 1077-1081 - João P. Cabral:
Uniform concatenative excitation model for synthesising speech without voiced/unvoiced classification. 1082-1086
Metadata, Evaluation and Resources I, II
- Matthias Sperber, Graham Neubig, Christian Fügen, Satoshi Nakamura, Alex Waibel:
Efficient speech transcription through respeaking. 1087-1091 - Samuel Kim, Panayiotis G. Georgiou, Shrikanth S. Narayanan:
Annotation and classification of Political advertisements. 1092-1096 - Ryuichiro Higashinaka, Kohji Dohsaka, Hideki Isozaki:
Using role play for collecting question-answer pairs for dialogue agents. 1097-1100 - Yoshiko Arimoto, Kazuo Okanoya:
Individual differences of emotional expression in speaker's behavioral and autonomic responses. 1101-1105 - Ina Wechsung, Benjamin Weiss, Christine Kühnel, Patrick Ehrenbrink, Sebastian Möller:
Development and validation of the conversational agents scale (CAS). 1106-1110 - Giuseppe Riccardi, Arindam Ghosh, S. A. Chowdhury, Ali Orkan Bayer:
Motivational feedback in crowdsourcing: a case study in speech transcription. 1111-1115 - Charles Fox, Yulan Liu, Erich Zwyssig, Thomas Hain:
The sheffield wargames corpus. 1116-1120 - Anuj Kumar, Florian Metze, Wenyi Wang, Matthew Kam:
Formalizing expert knowledge for developing accurate speech recognizers. 1121-1125 - Samer Al Moubayed, Jens Edlund, Joakim Gustafson:
Analysis of gaze and speech patterns in three-party quiz game interaction. 1126-1130 - Olivier Galibert:
Methodologies for the evaluation of speaker diarization and automatic speech recognition in the presence of overlapping speech. 1131-1134 - Abhijeet Sangwan, Lakshmish Kaushik, Chengzhu Yu, John H. L. Hansen, Douglas W. Oard:
'houston, we have a solution': using NASA apollo program to advance speech and language processing technology. 1135-1139
Speech Technology for Speech and Hearing Disorders I, II
- Robin Hofe, Jie Bai, Lam Aun Cheah, Stephen R. Ell, James M. Gilbert, Roger K. Moore, Phil D. Green:
Performance of the MVOCA silent speech interface across multiple speakers. 1140-1143 - Gustavo Andrade-Miranda, Juan Ignacio Godino-Llorente:
Automatic glottal tracking from high-speed digital images using a continuous normalized cross correlation. 1144-1148 - Tobias Bocklet, Stefan Steidl, Elmar Nöth, Sabine Skodda:
Automatic evaluation of parkinson's speech - acoustic, prosodic and voice related cues. 1149-1153 - Luiza Orosanu, Denis Jouvet:
Comparison of approaches for an efficient phonetic decoding. 1154-1158 - Heidi Christensen, Phil D. Green, Thomas Hain:
Learning speaker-specific pronunciations of disordered speech. 1159-1163 - Verónica López-Ludeña, Rubén San Segundo, Carlos González-Morcillo, Juan Carlos López, E. Ferreiro:
Adapting a speech into sign language translation system to a new domain. 1164-1168
Speech Analysis I-IV
- Elizabeth Godoy, Maria Koutsogiannaki, Yannis Stylianou:
Assessing the intelligibility impact of vowel space expansion via clear speech-inspired frequency warping. 1169-1173 - Jesper Jensen, Cees H. Taal:
Prediction of intelligibility of noisy and time-frequency weighted speech based on mutual information between amplitude envelopes. 1174-1178 - Emma Jokinen, Marko Takanen, Paavo Alku:
Frequency-adaptive post-filtering for intelligibility enhancement of narrowband telephone speech. 1179-1183 - Junfeng Li, Fei Chen, Masato Akagi, Yonghong Yan:
Comparative investigation of objective speech intelligibility prediction measures for noise-reduced signals in Mandarin and Japanese. 1184-1187 - Andrew Hines, Jan Skoglund, Anil C. Kokaram, Naomi Harte:
Monitoring the effects of temporal clipping on voIP speech quality. 1188-1192 - Jiahong Yuan:
The spectral dynamics of vowels in Mandarin Chinese. 1193-1197
Discriminative Training Methods for Language Modeling
- Holger Schwenk:
CSLM - a modular open-source continuous space language modeling toolkit. 1198-1202 - Yangyang Shi, Mei-Yuh Hwang, Kaisheng Yao, Martha A. Larson:
Speed up of recurrent neural network language models with sentence independent subsampling stochastic gradient descent. 1203-1207 - Shuangyu Chang, Michael Levit, Partha Parthasarathy, Benoît Dumoulin:
Improving unsupervised language model adaptation with discriminative data filtering. 1208-1212 - Akio Kobayashi, Takahiro Oku, Yuya Fujita, Shoei Sato:
Lightly supervised training for risk-based discriminative language models. 1213-1217 - Erinç Dikici, Emily Tucker Prud'hommeaux, Brian Roark, Murat Saraçlar:
Investigation of MT-based ASR confusion models for semi-supervised discriminative language modeling. 1218-1222 - Takanobu Oba, Atsunori Ogawa, Takaaki Hori, Hirokazu Masataki, Atsushi Nakamura:
Unsupervised discriminative language modeling using error rate estimator. 1223-1227
ASR - Adaptive Training
- Shakti P. Rath, Lukás Burget, Martin Karafiát, Ondrej Glembek, Jan Cernocký:
A region-specific feature-space transformation for speaker adaptation and singularity analysis of jacobian matrix. 1228-1232 - Yongqiang Wang, Mark J. F. Gales:
An explicit independence constraint for factorised adaptation in speech recognition. 1233-1237 - Oscar Saz, Thomas Hain:
Asynchronous factorisation of speaker and background with feature transforms in speech recognition. 1238-1242 - Kai Yu, Hainan Xu:
Cluster adaptive training with factorized decision trees for speech recognition. 1243-1247 - Ossama Abdel-Hamid, Hui Jiang:
Rapid and effective speaker adaptation of convolutional neural network based models for speech recognition. 1248-1252 - Keith Kintzley, Aren Jansen, Hynek Hermansky:
Text-to-speech inspired duration modeling for improved whole-word acoustic models. 1253-1257
Speech Acquisition and Development
- Adele Gregory, Marija Tabain, Michael Robb:
Duration of early vocalisations. 1258-1262 - Jing Yang, Robert Allen Fox:
Acoustic development of vowel production in American English children. 1263-1267 - Clément Moulin-Frier, Pierre-Yves Oudeyer:
The role of intrinsic motivations in learning sensorimotor vocal mappings: a developmental robotics study. 1268-1272 - Valérie Hazan, Michèle Pettinato:
Children's timing and repair strategies for communication in adverse listening conditions. 1273-1277 - Guillaume Barbier, Pascal Perrier, Lucie Ménard, Yohan Payan, Mark K. Tiede, Joseph S. Perkell:
Speech planning as an index of speech motor control maturity. 1278-1282 - Melissa Kinsman, Fangfang Li:
The relationship between gender-differentiated productions of /s/ and gender role behaviour in young children. 1283-1286
Articulatory Data Acquisition and Processing (Special Session)
- Jeffrey Berry, Luciano Fadiga:
Data-driven design of a sentence list for an articulatory speech corpus. 1287-1291 - Yinghua Zhu, Asterios Toutios, Shrikanth S. Narayanan, Krishna S. Nayak:
Faster 3d vocal tract real-time MRI using constrained reconstruction. 1292-1296 - Claudia Canevari, Leonardo Badino, Luciano Fadiga, Giorgio Metta:
Relevance-weighted-reconstruction of articulatory features in deep-neural-network-based acoustic-to-articulatory mapping. 1297-1301 - Fabian Tomaschek, Martijn Wieling, Denis Arnold, R. Harald Baayen:
Word frequency, vowel length and vowel quality in speech production: an EMA study of the importance of experience. 1302-1306 - Samuel S. Silva, António J. S. Teixeira, Catarina Oliveira, Paula Martins:
Towards a systematic and quantitative analysis of vocal tract data. 1307-1311 - Colin Vaz, Vikram Ramanarayanan, Shrikanth S. Narayanan:
A two-step technique for MRI audio enhancement using dictionary learning and wavelet packet analysis. 1312-1315 - Massimo Stella, Antonio Stella, Francesco Sigona, Paolo Bernardini, Mirko Grimaldi, Barbara Gili Fivela:
Electromagnetic articulography with AG500 and AG501. 1316-1320 - Pierre Badin, Julián Andrés Valdés Vargas, Arielle Koncki, Laurent Lamalle, Christophe Savariaux:
Development and implementation of fiducial markers for vocal tract MRI imaging and speech articulatory modelling. 1321-1325 - Susanne Schötz, Johan Frid, Lars Gustafsson, Anders Löfqvist:
Functional data analysis of tongue articulation in palatal vowels: gothenburg and malmöhus Swedish /iː, yː, ̟ʉː/. 1326-1330 - Jordan R. Green, Jun Wang, David L. Wilson:
SMASH: a tool for articulatory data processing and analysis. 1331-1335
Topics in Speech Perception and Emotion
- Jen-Chun Lin, Chung-Hsien Wu, Wen-Li Wei:
Emotion recognition of conversational affective speech using temporal course modeling. 1336-1340 - Rene Altrov, Hille Pajupuu, Jaan Pajupuu:
The role of empathy in the recognition of vocal emotions. 1341-1344 - Angèle Brunellière, Sophie Dufour:
Electrophysiological evidence for benefits of imitation during the processing of spoken words embedded in sentential contexts. 1345-1349 - Rintaro Ogane, Masaaki Honda:
Compensatory speech response to time-scale altered auditory feedback. 1350-1354 - Tin Lay Nwe, Trung Hieu Nguyen, Dilip Kumar Limbu:
Bhattacharyya distance based emotional dissimilarity measure in multi-dimensional space for emotion classification. 1355-1359 - Thiago de M. Prego, Amaro A. de Lima, Sergio L. Netto:
On the enhancement of dereverberation algorithms based on a perceptual evaluation criterion. 1360-1364 - Carlos Gussenhoven, Wencui Zhou:
Revisiting pitch slope and height effects on perceived duration. 1365-1369 - Hélène Guiraud, Emmanuel Ferragne, Nathalie Bedoin, Véronique Boulenger:
Adaptation to natural fast speech and time-compressed speech in children. 1370-1374 - Andreas Windmann, Juraj Simko, Britta Wrede, Petra Wagner:
Modeling durational incompressibility. 1375-1379 - Caroline Émond, Lucie Ménard, Marty Laforest:
Perceived prosodic correlates of smiled speech in spontaneous data. 1380-1383 - Alexander Raake, Katrin Schoenenberg, Janto Skowronek, Sebastian Egger:
Predicting speech quality based on interactivity and delay. 1384-1388 - Charlotte Kouklia, Nicolas Audibert:
Perceptual, acoustic and electroglottographic correlates of 3 aggressive attitudes in French: a pilot study. 1389-1393
Discourse and Machine Learning, Paralinguistic and Nonlinguistic Cues
- Mohamed Morchid, Georges Linarès, Marc El-Bèze, Renato De Mori:
Theme identification in telephone service conversations using quaternions of speech features. 1394-1398 - Hrishikesh Rao, Jonathan C. Kim, Agata Rozga, Mark A. Clements:
Detection of laughter in children's speech using spectral and prosodic acoustic features. 1399-1403 - Khiet P. Truong:
Classification of cooperative and competitive overlaps in speech using cues from the context, overlapper, and overlappee. 1404-1408 - Samuel Kim, Fabio Valente, Alessandro Vinciarelli:
Annotation and detection of conflict escalation in Political debates. 1409-1413 - Florian Schiel, Mary Stevens, Uwe D. Reichel, Francesco Cutugno:
Machine learning of probabilistic phonological pronunciation rules from the Italian CLIPS corpus. 1414-1418 - Barbara Baumeister, Florian Schiel:
Human perception of alcoholic intoxication in speech. 1419-1423 - Luying Hou, Yuan Jia, Aijun Li:
Phonetic manifestation and influence of zero anaphora in Chinese reading texts. 1424-1428 - Salima Harrat, Mourad Abbas, Karima Meftouh, Kamel Smaïli:
Diacritics restoration for Arabic dialect texts. 1429-1433 - Marcin Wlodarczak, Petra Wagner:
Effects of talk-spurt silence boundary thresholds on distribution of gaps and overlaps. 1434-1437 - Tatiana Kachkovskaia, Nina B. Volskaya, Pavel A. Skrelin:
Final lengthening in Russian: a corpus-based study. 1438-1442 - Uwe D. Reichel:
From segmentation bootstrapping to transcription-to-word conversion. 1443-1447 - Geneviève Caelen-Haumont, Katarina Bartkova:
Manual and automatic tone annotation: the case of an endangered language from north vietnam "mo piu". 1448-1452 - Laetitia Leonarduzzi, Sophie Herment:
Non-canonical syntactic structures in discourse: tonality, tonicity and tones in English (semi-)spontaneous speech. 1453-1457 - Elnaz Nouri, Sunghyun Park, Stefan Scherer, Jonathan Gratch, Peter J. Carnevale, Louis-Philippe Morency, David R. Traum:
Prediction of strategy and outcome as negotiation unfolds by using basic verbal and behavioral features. 1458-1461
Language Identification, Speaker Diarization
- Johann Poignant, Laurent Besacier, Viet Bac Le, Sophie Rosset, Georges Quénot:
Unsupervised naming of speakers in broadcast TV: using written names, pronounced names or both? 1462-1466 - Hervé Bredin, Johann Poignant:
Integer linear programming for speaker diarization and cross-modal identification in TV broadcast. 1467-1471 - Andrea DeMarco, Stephen J. Cox:
Native accent classification via i-vectors and speaker compensation fusion. 1472-1476 - Mickael Rouvier, Grégor Dupuy, Paul Gay, Elie Khoury, Téva Merlin, Sylvain Meignier:
An open-source state-of-the-art toolbox for broadcast news diarization. 1477-1481 - Zvi Kons, Orith Toledo-Ronen:
Audio event classification using deep neural networks. 1482-1486 - Wei-Bin Liang, Chung-Hsien Wu, Chun-Shan Hsu:
Code-Switching event detection based on delta-BIC using phonetic eigenvoice models. 1487-1491 - Naoki Hirayama, Koichiro Yoshino, Katsutoshi Itoyama, Shinsuke Mori, Hiroshi G. Okuno:
Automatic estimation of dialect mixing ratio for dialect speech recognition. 1492-1496 - Luis Javier Rodríguez-Fuentes, Niko Brümmer, Mikel Peñagarikano, Amparo Varona, Germán Bordel, Mireia Díez:
The albayzin 2012 language recognition evaluation. 1497-1501 - Kyu Jeong Han, Sriram Ganapathy, Ming Li, Mohamed Kamal Omar, Shrikanth S. Narayanan:
TRAP language identification system for RATS phase II evaluation. 1502-1506 - Aaron Lawson, Mitchell McLaren, Yun Lei, Vikramjit Mitra, Nicolas Scheffer, Luciana Ferrer, Martin Graciarena:
Improving language identification robustness to highly channel-degraded speech through multiple system fusion. 1507-1510
Metadata, Evaluation and Resources I, II
- Jindrich Matousek, Daniel Tihelka:
Annotation errors detection in TTS corpora. 1511-1515 - Imran Ahmed, Sunil Kumar Kopparapu:
Technique for automatic sentence level alignment of long speech and transcripts. 1516-1519 - Sarah Hoffmann, Beat Pfister:
Text-to-speech alignment of long recordings using universal phone models. 1520-1524 - Adriana Stan, Peter Bell, Junichi Yamagishi, Simon King:
Lightly supervised discriminative training of grapheme models for improved sentence-level alignment of speech and text data. 1525-1529 - Ashtosh Sapru, Hervé Bourlard:
Automatic social role recognition in professional meetings using conditional random fields. 1530-1534 - Christoph Draxler, Hanna S. Feiser:
Same same but different - an acoustical comparison of the automatic segmentation of high quality and mobile telephone speech. 1535-1539
Speech Synthesis - Prosody and Emotion
- Yongguo Kang, Jian Li, Yan Deng, Miaomiao Wang:
Multi-centroidal duration generation algorithm for HMM-based TTS. 1540-1543 - Tuomo Raitio, Antti Suni, Jouni Pohjalainen, Manu Airaksinen, Martti Vainio, Paavo Alku:
Analysis and synthesis of shouted speech. 1544-1548 - Tomohiro Nagata, Hiroki Mori, Takashi Nose:
Robust estimation of multiple-regression HMM parameters for dimension-based expressive dialogue speech synthesis. 1549-1553 - Sandrine Brognaux, Benjamin Picart, Thomas Drugman:
A new prosody annotation protocol for live sports commentaries. 1554-1558 - Mahnoosh Mehrabani, Taniya Mishra, Alistair Conkie:
Unsupervised prominence prediction for speech synthesis. 1559-1563 - Marcela Charfuelan, Ingmar Steiner:
Expressive speech synthesis in MARY TTS using audiobook data and emotionML. 1564-1568
Spoken Language Information Retrieval
- Nigel G. Ward, Steven D. Werner:
Using dialog-activity similarity for spoken information retrieval. 1569-1573 - I-Fan Chen, Chin-Hui Lee:
A hybrid HMM/DNN approach to keyword spotting of short words. 1574-1578 - Jonathan Wintrode:
Leveraging locality for topic identification of conversational speech. 1579-1583 - Grégory Senay, Benjamin Bigot, Richard Dufour, Georges Linarès, Corinne Fredouille:
Person name spotting by combining acoustic matching and LDA topic models. 1584-1588 - György Szaszák, András Beke:
Using phonological phrase segmentation to improve automatic keyword spotting for the highly agglutinating Hungarian language. 1589-1593 - Larry P. Heck, Dilek Hakkani-Tür, Gökhan Tür:
Leveraging knowledge graphs for web-scale unsupervised semantic parsing. 1594-1598
Speaker Recognition I, II
- Sandro Cumani, Pietro Laface:
Fast and memory effective i-vector extraction using a factorized sub-space. 1599-1603 - Konstantin Simonchik, Andrey Shulipa, Timur Pekhovsky:
Effective estimation of a multi-session speaker model using information on signal parameters. 1604-1608 - Ville Hautamäki, Kong-Aik Lee, David A. van Leeuwen, Rahim Saeidi, Anthony Larcher, Tomi Kinnunen, Taufiq Hasan, Seyed Omid Sadjadi, Gang Liu, Hynek Boril, John H. L. Hansen, Benoit G. B. Fauve:
Automatic regularization of cross-entropy cost for speaker recognition fusion. 1609-1613 - Ming Li, Jangwon Kim, Prasanta Kumar Ghosh, Vikram Ramanarayanan, Shrikanth S. Narayanan:
Speaker verification based on fusion of acoustic and articulatory information. 1614-1618 - David A. van Leeuwen, Niko Brümmer:
The distribution of calibrated likelihood-ratios in speaker recognition. 1619-1623 - Finnian Kelly, Niko Brümmer, Naomi Harte:
Eigenageing compensation for speaker verification. 1624-1628
Multimodal Speech Perception
- Grozdana Erjavec, Denis Legros:
Effects of mouth-only and whole-face displays on audio-visual speech perception in noise: is the vision of a talker's full face truly the most efficient solution? 1629-1633 - Kaisa Tiippana, Mikko Tiainen, Lari Vainio, Martti Vainio:
Acoustic and visual phonetic features in the mcgurk effect - an audiovisual speech illusion. 1634-1638 - Chris Davis, Jeesun Kim:
The effect of visual speech timing and form cues on the processing of speech and nonspeech. 1639-1642 - Ganesh Attigodu Chandrashekara, Frédéric Berthommier, Olha Nahorna, Jean-Luc Schwartz:
Effect of context, rebinding and noise, on audiovisual speech fusion. 1643-1647 - Albert Rilliard, Donna Erickson, Takaaki Shochi, João Antônio de Moraes:
Social face to face communication - American English attitudinal prosody. 1648-1652 - Gérard Bailly, Amélie Rochet-Capellan, Coriandre Vilain:
Adaptation of respiratory patterns in collaborative reading. 1653-1657
Speech Analysis I-IV
- John Kane, Stefan Scherer, Louis-Philippe Morency, Christer Gobl:
A comparative study of glottal open quotient estimation techniques. 1658-1662 - Christian H. Kasess, Wolfgang Kreuzer:
Estimation of multiple-branch vocal tract models: the influence of prior assumptions. 1663-1667 - Jürgen T. Geiger, Florian Eyben, Björn W. Schuller, Gerhard Rigoll:
Detecting overlapping speech with long short-term memory recurrent neural networks. 1668-1672 - Akira Sasou:
Evaluation of fundamental validity in applying AR-HMM with automatic topology generation to pathology voice analysis. 1673-1676 - Nagaraj Adiga, S. R. M. Prasanna:
Significance of instants of significant excitation for source modeling. 1677-1681 - Devanshu Arya, Anant Raj, Rajesh M. Hegde:
Significance of variable height-bandwidth group delay filters in the spectral reconstruction of speech. 1682-1686 - Hemant A. Patil, Tanvina B. Patel:
Nonlinear prediction of speech signal using volterra-wiener series. 1687-1691 - Aharon Satt, Alexander Sorin, Orith Toledo-Ronen, Oren Barkan, Ioannis Kompatsiaris, Athina Kokonozi, Magda Tsolaki:
Evaluation of speech-based protocol for detection of early-stage dementia. 1692-1696 - Elias Azarov, Maxim Vashkevich, Alexander A. Petrovsky:
Instantaneous harmonic representation of speech using multicomponent sinusoidal excitation. 1697-1701 - Onur Babacan, Thomas Drugman, Nicolas D'Alessandro, Nathalie Henrich, Thierry Dutoit:
A quantitative comparison of glottal closure instant estimation algorithms on a large variety of singing sounds. 1702-1706 - Jorge Andrés Gómez García, Juan Ignacio Godino-Llorente, Germán Castellanos-Domínguez:
Automatic gender recognition in normal and pathological speech. 1707-1711 - Shanqing Cai, H. Timothy Bunnell, Rupal Patel:
Unsupervised vocal-tract length estimation through model-based acoustic-to-articulatory inversion. 1712-1716 - Sayeh Mirzaei, Hugo Van hamme, Yaser Norouzi:
Model order estimation using Bayesian NMF for discovering phone patterns in spoken utterances. 1717-1721
ASR - Feature Extraction
- László Tóth:
Convolutional deep rectifier neural nets for phone recognition. 1722-1726 - Hans-Günter Hirsch:
Pitch synchronous spectral analysis for a pitch dependent recognition of voiced phonemes - PISAR. 1727-1731 - José Luis Oropeza Rodríguez:
New parameters for automatic speech recognition based on the mammalian cochlea model using resonance analysis. 1732-1736 - Navdeep Jaitly, Geoffrey E. Hinton:
Using an autoencoder with deformable templates to discover features for automated speech recognition. 1737-1740 - Ching-feng Yeh, Hung-yi Lee, Lin-Shan Lee:
Speaking rate normalization with lattice-based context-dependent phoneme duration modeling for personalized speech recognizers on mobile devices. 1741-1745 - Jun Qi, Dong Wang, Javier Tejedor:
Subspace models for bottleneck features. 1746-1750 - Jun Qi, Dong Wang, Ji Xu, Javier Tejedor:
Bottleneck features based on gammatone frequency cepstral coefficients. 1751-1755 - Pavel Golik, Patrick Doetsch, Hermann Ney:
Cross-entropy vs. squared error training: a theoretical and experimental comparison. 1756-1760 - Vaishali Patil, Preeti Rao:
Acoustic features for detection of phonemic aspiration in voiced plosives. 1761-1765 - Dimitri Palaz, Ronan Collobert, Mathew Magimai-Doss:
Estimating phoneme class conditional probabilities from raw speech signal using convolutional neural networks. 1766-1770 - Javier Mikel Olaso, M. Inés Torres:
Hierarchical models based on a continuous acoustic space to identify phonological features. 1771-1775 - Vikrant Singh Tomar, Richard C. Rose:
Locality sensitive hashing for fast computation of correlational manifold learning based feature space transformations. 1776-1780 - Thomas Schatz, Vijayaditya Peddinti, Francis R. Bach, Aren Jansen, Hynek Hermansky, Emmanuel Dupoux:
Evaluating speech features with the minimal-pair ABX task: analysis of the classical MFC/PLP pipeline. 1781-1785
ASR - Pronunciation, Prosodic and New Paradigms
- Chen-Yu Chiang, Sabato Marco Siniscalchi, Sin-Horng Chen, Chin-Hui Lee:
Knowledge integration for improving performance in LVCSR. 1786-1790 - Martin Heckmann:
Inter-speaker variability in audio-visual classification of word prominence. 1791-1795 - Shilin Liu, Khe Chai Sim:
Parameter clustering for temporally varying weight regression for automatic speech recognition. 1796-1800 - Tanel Alumäe, Rena Nemoto:
Phone duration modeling using clustering of rich contexts. 1801-1805 - Farzaneh Ahmadi, Mousa Ahmadi, Ian McLoughlin:
Human mouth state detection using low frequency ultrasound. 1806-1810 - Kun Li, Xiaojun Qian, Shiyin Kang, Helen Meng:
Lexical stress detection for L2 English speech using deep belief networks. 1811-1815 - Yanmin Qian, Jia Liu:
MLP-HMM two-stage unsupervised training for low-resource languages on conversational telephone speech recognition. 1816-1820 - Josef R. Novak, Nobuaki Minematsu, Keikichi Hirose:
Failure transitions for joint n-gram models and G2p conversion. 1821-1825 - Hirokazu Kameoka, Kota Yoshizato, Tatsuma Ishihara, Yasunori Ohishi, Kunio Kashino, Shigeki Sagayama:
Generative modeling of speech F0 contours. 1826-1830 - Marelie H. Davel, Charl Johannes van Heerden, Etienne Barnard:
G2p variant prediction techniques for ASR and STD. 1831-1835 - Jin Jin, Joseph Tepperman:
Rhythm analysis of second-language speech through low-frequency auditory features. 1836-1839 - Yuzong Liu, Katrin Kirchhoff:
Graph-based semi-supervised learning for phone and segment classification. 1840-1843 - Ao Shen, Neil Cooke, Martin J. Russell:
Selective use of gaze information to improve ASR performance in noisy environments by cache-based class language model adaptation. 1844-1848 - Ossama Abdel-Hamid, Li Deng, Dong Yu, Hui Jiang:
Deep segmental neural networks for speech recognition. 1849-1853 - Martine Coene, Annemiek Hammer, Wojtek Kowalczyk, Louis ten Bosch, Bart Vaerenberg, Paul Govaerts:
Quantifying cross-linguistic variation in grapheme-to-phoneme mapping. 1854-1857
Show and Tell Sessions 1-3
- Florian Metze, Eric Fosler-Lussier, Rebecca Bates:
The speech recognition virtual kitchen. 1858-1860 - John Chen, Shufei Wen, Vivek Kumar Rangarajan Sridhar, Srinivas Bangalore:
Multilingual web conferencing using speech-to-speech translation. 1861-1863 - Emmanuel Ferragne, Sébastien Flavier, Christian Fressard:
ROCme! software for the recording and management of speech corpora. 1864-1865 - Felix Burkhardt:
Voice search in mobile applications with the rootvole framework. 1866-1868 - John S. Novak III, Jason Archer, Valeriy Shafiro, Robert V. Kenyon, Jason Leigh:
On-line audio dilation for human interaction. 1869-1871 - Pejman Mowlaee, Mario Kaoru Watanabe, Rahim Saeidi:
Phase-aware single-channel speech enhancement. 1872-1874 - Hiroko Hirano, Ibuki Nakamura, Nobuaki Minematsu, Masayuki Suzuki, Chieko Nakagawa, Noriko Nakamura, Yukinori Tagawa, Keikichi Hirose, Hiroya Hashimoto:
A free online accent and intonation dictionary for teachers and learners of Japanese. 1875-1876 - Maria Astrinaki, Junichi Yamagishi, Simon King, Nicolas D'Alessandro, Thierry Dutoit:
Reactive accent interpolation through an interactive map application. 1877-1878 - Kay Berkling:
A non-experts user interface for obtaining automatic diagnostic spelling evaluations for learners of the German writing system. 1879-1881
Dialog Systems
- Tatsuya Kawahara, Soichiro Hayashi, Katsuya Takanashi:
Estimation of interest and comprehension level of audience through multi-modal behaviors in poster conversations. 1882-1885 - Wenping Hu, Yao Qian, Frank K. Soong:
A new DNN-based high quality pronunciation evaluation for computer-aided language learning (CALL). 1886-1890 - Joaquin Planells, Lluís F. Hurtado, Encarna Segarra, Emilio Sanchis:
A multi-domain dialog system to integrate heterogeneous spoken dialog systems. 1891-1895 - Yuki Todo, Ryota Nishimura, Kazumasa Yamamoto, Seiichi Nakagawa:
Development and evaluation of spoken dialog systems with one or two agents. 1896-1900 - Gabriel Skantze, Catharine Oertel, Anna Hjalmarsson:
User feedback in human-robot interaction: prosody, gaze and timing. 1901-1905 - Yongxin Taylor Xi, Matthias Paulik, Venkata Ramana Rao Gadde, Ananth Sankar:
KPCatcher - a keyphrase extraction system for enterprise videos. 1906-1910
Speech Analysis I-IV
- Malcolm Slaney, Elizabeth Shriberg, Jui-Ting Huang:
Pitch-gesture modeling using subband autocorrelation change detection. 1911-1915 - P. Gangamohan, Sudarsana Reddy Kadiri, B. Yegnanarayana:
Analysis of emotional speech at subsegmental level. 1916-1920 - Masanori Morise, Hideki Kawahara, Kenji Ozawa:
Periodicity extraction for voiced sounds with multiple periodicity. 1921-1925 - John H. Taylor, Ben Milner:
Modelling and estimation of the fundamental frequency of speech using a hidden Markov model. 1926-1930 - Jouni Pohjalainen, Paavo Alku:
Extended weighted linear prediction using the autocorrelation snapshot - a robust speech analysis method and its application to recognition of vocal emotions. 1931-1935 - Meysam Asgari, Izhak Shafran:
Improving the accuracy and the robustness of harmonic model for pitch estimation. 1936-1940
ASR - Pronunciation Variants and Modeling
- Meixu Song, Qingqing Zhang, Jielin Pan, Yonghong Yan:
Discriminative pronunciation modeling based on minimum phone error training. 1941-1945 - Keigo Kubo, Sakriani Sakti, Graham Neubig, Tomoki Toda, Satoshi Nakamura:
Grapheme-to-phoneme conversion based on adaptive regularization of weight vectors. 1946-1950 - Tofigh Naghibi, Sarah Hoffmann, Beat Pfister:
An efficient method to estimate pronunciation from multiple utterances. 1951-1955 - Willem D. Basson, Marelie H. Davel:
Category-based phoneme-to-grapheme transliteration. 1956-1960 - Preethi Jyothi, Eric Fosler-Lussier, Karen Livescu:
Discriminative training of WFST factors with application to pronunciation modeling. 1961-1965 - Penny Karanasou, François Yvon, Thomas Lavergne, Lori Lamel:
Discriminative training of a phoneme confusion model for a dynamic lexicon in ASR. 1966-1970
Speaker Recognition Evaluation
- Craig S. Greenberg, Vincent M. Stanford, Alvin F. Martin, Meghana Yadagiri, George R. Doddington, John J. Godfrey, Jaime Hernandez-Cordero:
The 2012 NIST speaker recognition evaluation. 1971-1975 - Niko Brümmer, George R. Doddington:
Likelihood-ratio calibration using prior-weighted proper scoring rules. 1976-1980 - Luciana Ferrer, Mitchell McLaren, Nicolas Scheffer, Yun Lei, Martin Graciarena, Vikramjit Mitra:
A noise-robust system for NIST 2012 speaker recognition evaluation. 1981-1985 - Rahim Saeidi, Kong-Aik Lee, Tomi Kinnunen, Tawfik Hasan, Benoit G. B. Fauve, Pierre-Michel Bousquet, Elie Khoury, Pablo Luis Sordo Martinez, Jia Min Karen Kua, Changhuai You, Hanwu Sun, Anthony Larcher, Padmanabhan Rajan, Ville Hautamäki, Cemal Hanilçi, Billy Braithwaite, Rosa González Hautamäki, Seyed Omid Sadjadi, Gang Liu, Hynek Boril, Navid Shokouhi, Driss Matrouf, Laurent El Shafey, Pejman Mowlaee, Julien Epps, Tharmarajah Thiruvaran, David A. van Leeuwen, Bin Ma, Haizhou Li, John H. L. Hansen, Jean-François Bonastre, Sébastien Marcel, John S. D. Mason, Eliathamby Ambikairajah:
I4u submission to NIST SRE 2012: a large-scale collaborative effort for noise-robust speaker verification. 1986-1990 - Hanwu Sun, Bin Ma:
Improved unsupervised NAP training dataset design for speaker recognition. 1991-1995 - Daniele Colibro, Claudio Vair, Kevin Farrell, Nir Krause, Gennady Karvitsky, Sandro Cumani, Pietro Laface:
Nuance - Politecnico di torino's 2012 NIST speaker recognition evaluation system. 1996-2000
Physiology and Models of Speech Production
- Gang Chen, Marc Garellek, Jody Kreiman, Bruce R. Gerratt, Abeer Alwan:
A perceptually and physiologically motivated voice source model. 2001-2005 - Caitlin Smith, Michael I. Proctor, Khalil Iskarous, Louis Goldstein, Shrikanth S. Narayanan:
Stable articulatory tasks and their variable formation: tamil retroflex consonants. 2006-2009 - Vikram Ramanarayanan, Adam C. Lammert, Louis Goldstein, Shrikanth S. Narayanan:
Articulatory settings facilitate mechanically advantageous motor control of vocal tract articulators. 2010-2013 - Amélie Rochet-Capellan, Susanne Fuchs:
The interplay of linguistic structure and breathing in German spontaneous speech. 2014-2018 - Takayuki Arai:
Physical models of the vocal tract with a flapping tongue for flap and liquid sounds. 2019-2023 - Yves Laprie, Matthieu Loosvelt, Shinji Maeda, Rudolph Sock, Fabrice Hirsch:
Articulatory copy synthesis from cine x-ray films. 2024-2028
Speech Science in End-User Applications
- Jerome R. Bellegarda:
Large-scale personal assistant technology deployment: the siri experience. 2029-2033 - Benjamin Weiss, Simon Willkomm, Sebastian Möller:
Evaluating an adaptive dialog system for the public. 2034-2038 - Jort F. Gemmeke, Bart Ons, Netsanet M. Tessema, Hugo Van hamme, Janneke van de Loo, Guy De Pauw, Walter Daelemans, Jonathan Huyghe, Jan Derboven, Lode Vuegen, Bert Van Den Broeck, Peter Karsmakers, Bart Vanrumste:
Self-taught assistive vocal interfaces: an overview of the ALADIN project. 2039-2043 - Florian Eyben, Felix Weninger, Björn W. Schuller:
Affect recognition in real-life acoustic conditions - a new perspective on feature selection. 2044-2048 - Emanuele Principi, Stefano Squartini, Francesco Piazza, Danilo Fuselli, Maurizio Bonifazi:
A distributed system for recognizing home automation commands and distress calls in the Italian language. 2049-2053 - Nina Zinovieva, Xiaodan Zhuang, Pat Peterson, Joe Alwan, Rohit Prasad:
Probabilistic trainable segmenter for call center audio using multiple features. 2054-2058 - Felix Burkhardt, Hans Ulrich Nägeli:
Voice search in mobile applications and the use of linked open data. 2059-2061 - Michel Vacher, Benjamin Lecouteux, Dan Istrate, Thierry Joubert, François Portet, Mohamed A. Sehili, Pedro Chahuara:
Evaluation of a real-time voice order recognition system from multiple audio channels in a home. 2062-2064 - Frédéric Aman, Michel Vacher, Solange Rossato, François Portet:
In-home detection of distress calls: the case of aged users. 2065-2067 - Ding Liu, Anthea Cheung, Anna Margolis, Patrick Redmond, Jun-Won Suh, Chao Wang:
Data driven methods for utterance semantic tagging. 2068-2070 - Evandro Gouvêa, Antonio Moreno-Daniel, A. Reddy, Rathinavelu Chengalvarayan, David L. Thomson, Andrej Ljolje:
The AT&t speech API: a study on practical challenges for customized speech to text service. 2071-2073 - Bart D'hoore, Alfred Wiesen:
In-vehicle destination entry by voice: practical aspects. 2074-2076
Perception of Non Native Sounds
- Aurore Gautreau, Michel Hoen, Fanny Meunier:
Intelligibility at a multilingual cocktail party: effect of concurrent language knowledge. 2077-2080 - Ewa Jacewicz, Robert Allen Fox:
Regional accents affect speech intelligibility in a multitalker environment. 2081-2085 - Shinichi Tokuma, Won Tokuma:
Perception of English minimal pairs in noise by Japanese listeners: does clear speech for L2 listeners help? 2086-2090 - Bianca Sisinni, Paola Escudero, Mirko Grimaldi:
Salento Italian listeners' perception of American English vowels. 2091-2094 - Andréia Schurt Rauber, Anabela Rato, Denise Cristina Kluge, Giane Rodrigues dos Santos:
TP 3.1 software: a tool for designing audio, visual, and audiovisual perceptual training tasks and perception tests. 2095-2098 - Fei Chen, Junfeng Li, Lena L. N. Wong, Yonghong Yan:
Effect of linguistic masker on the intelligibility of Mandarin sentences. 2099-2102 - Kyuwon Moon, Meghan Sumner:
The learning and generalization of contrasts consistent or inconsistent with native biases. 2103-2107 - Jia Ying, Jason A. Shaw, Catherine T. Best:
L2 English learners' recognition of words spoken in familiar versus unfamiliar English accents. 2108-2112 - Janice Wing Sze Wong:
The effects of perceptual and/or productive training on the perception and production of English vowels /ɪ/ and /iː/ by Cantonese ESL learners. 2113-2117 - Natalia Kartushina, Ulrich H. Frauenfelder:
On the role of L1 speech production in L2 perception: evidence from Spanish learners of French. 2118-2122 - Pierre A. Hallé, Natalia Kartushina, Juan Segui, Ulrich H. Frauenfelder:
Looking for lexical feedback effects in /tl/→/kl/ repairs. 2123-2127 - Catherine T. Best, Jason A. Shaw, Elizabeth Clancy:
Recognizing words across regional accents: the role of perceptual assimilation in lexical competition. 2128-2132
Speech Disorders - Data and Methodology
- David Martínez González, Phil D. Green, Heidi Christensen:
Dysarthria intelligibility assessment in a factor analysis total variability space. 2133-2137 - Alain Ghio, Médéric Gasquet-Cyrus, Juliette Roquel, Antoine Giovanni:
Perceptual interference between regional accent and voice/speech disorders. 2138-2142 - Ingrida Balciuniene:
Linguistic disfluency in narrative speech: evidence from story-telling in 6-year olds. 2143-2146 - Benjamin Munson:
Assessing the utility of judgments of children's speech production made by untrained listeners in uncontrolled listening environments. 2147-2151 - Tanja Kocjancic Antolík, Cécile Fougeron:
Consonant distortions in dysarthria due to parkinson's disease, amyotrophic lateral sclerosis and cerebellar ataxia. 2152-2156 - Marine Verdurand, Solange Rossato, Lionel Granjon, Daria Balbo, Claudio Zmarich:
Study of coarticulation and F2 transitions in French and Italian adult stutterers. 2157-2161 - Renee Peje Clapham, Corina J. van As-Brooks, Michiel W. M. van den Brekel, Frans J. M. Hilgers, R. J. J. H. van Son:
Automatic tracheoesophageal voice typing using acoustic parameters. 2162-2166 - Julie Mauclair, Lionel Koenig, Marina Robert, Peggy Gatignol:
Burst-based features for the classification of pathological voices. 2167-2171 - Brian S. Helfer, Thomas F. Quatieri, James R. Williamson, Daryush D. Mehta, Rachelle Horwitz, Bea Yu:
Classification of depression state based on articulatory precision. 2172-2176 - Kathleen C. Fraser, Frank Rudzicz, Elizabeth Rochon:
Using text and acoustic features to diagnose progressive aphasia and its subtypes. 2177-2181
Search and Computational Issues in LVCSR
- Tanel Alumäe:
Multi-domain neural network language model. 2182-2186 - Yanhua Long, Mark J. F. Gales, Pierre Lanchantin, Xunying Liu, Matthew Stephen Seigel, Philip C. Woodland:
Improving lightly supervised training for broadcast transcription. 2187-2191 - Christophe Cerisara, Alejandra Lorenzo, Pavel Král:
Weakly supervised parsing with rules. 2192-2196 - Markus Nußbaum-Thom, Eugen Beck, Tamer Alkhouli, Ralf Schlüter, Hermann Ney:
Relative error bounds for statistical classifiers based on the f-divergence. 2197-2201 - Melvin Jose Johnson Premkumar, Ngoc Thang Vu, Tanja Schultz:
Experiments towards a better LVCSR system for tamil. 2202-2206 - Kwanchiva Thangthai, Ananlada Chotimongkol, Chai Wutiwiwatchai:
A hybrid language model for open-vocabulary Thai LVCSR. 2207-2211 - Jen-Tzung Chien, Ying-Lan Chang:
Hierarchical pitman-yor and dirichlet process for language model. 2212-2216 - Taichi Asami, Satoshi Kobashikawa, Hirokazu Masataki, Osamu Yoshioka, Satoshi Takahashi:
Unsupervised confidence calibration using examples of recognized words and their contexts. 2217-2221 - Zoltán Tüske, Ralf Schlüter, Hermann Ney:
Multilingual hierarchical MRASTA features for ASR. 2222-2226 - Harry M. Chang:
Heuristic selection of training sentences from historical TV guide for semi-supervised LM adaptation. 2227-2231 - Dominique Fohr, Odile Mella:
Combination of random indexing based language model and n-gram language model for speech recognition. 2232-2236 - Yajie Miao, Florian Metze:
Improving low-resource CD-DNN-HMM using dropout and multilingual DNN training. 2237-2241 - Long Qin, Alexander I. Rudnicky:
Finding recurrent out-of-vocabulary words. 2242-2246 - Justin T. Chiu, Alexander I. Rudnicky:
Using conversational word bursts in spoken term detection. 2247-2251
Speech and Hearing Disorders
- Audrey Acher, Marc Sato, Laurent Lamalle, Coriandre Vilain, Arnaud Attye, Alexandre Krainik, Georges Bettega, Christian Adrien Righini, Brice Carlot, Muriel Brix, Pascal Perrier:
Brain activations in speech recovery process after intra-oral surgery: an fMRI study. 2252-2256 - Christophe Mertens, Jean Schoentgen, Francis Grenez, Sabine Skodda:
Acoustic and perceptual analysis of vocal tremor. 2257-2261 - Charturong Tantibundhit, Chutamanee Onsuwan, Nittayapa Klangpornkun, P. Phienphanich, Tanawan Saimai, Nantaporn Saimai, P. Pitathawatchai, Chai Wutiwiwatchai:
Lexical tone perception in Thai normal-hearing adults and those using hearing aids: a case study. 2262-2266 - Takayuki Kagomiya, Seiji Nakagawa:
Evaluation of a bone-conducted ultrasonic hearing aid in vocal emotion transmission. 2267-2271 - Luigia Garrapa, Davide Bottari, Mirko Grimaldi, Francesco Pavani, Andrea Calabrese, Michele De Benedetto, Silvano Vitale:
Processing of /i/ and /u/ in Italian cochlear-implant children: a behavioral and neurophysiologic study. 2272-2276 - Stefano Cosentino, Tiago H. Falk, David McAlpine:
Predicting the bilateral advantage in cochlear implantees using a non-intrusive speech intelligibility measure. 2277-2281
Speech and Audio Segmentation
- Zhen Huang, You-Chi Cheng, Kehuang Li, Ville Hautamäki, Chin-Hui Lee:
A blind segmentation approach to acoustic event detection based on i-vector. 2282-2286 - Van Zyl van Vuuren, Louis ten Bosch, Thomas Niesler:
A dynamic programming framework for neural network-based automatic speech segmentation. 2287-2291 - RaviShankar Prasad, B. Yegnanarayana:
Acoustic segmentation of speech using zero time liftering (ZTL). 2292-2296 - Haipeng Wang, Tan Lee, Cheung-Chi Leung, Bin Ma, Haizhou Li:
Unsupervised mining of acoustic subword units with segment-level Gaussian posteriorgrams. 2297-2301 - Ozlem Kalinli:
Combination of auditory attention features with phone posteriors for better automatic phoneme segmentation. 2302-2305 - Jiahong Yuan, Neville Ryant, Mark Liberman, Andreas Stolcke, Vikramjit Mitra, Wen Wang:
Automatic phonetic segmentation using boundary models. 2306-2310
Speech Synthesis - Various Topics
- Thi Thu Trang Nguyen, Christophe d'Alessandro, Albert Rilliard, Do Dat Tran:
HMM-based TTS for hanoi vietnamese: issues in design and evaluation. 2311-2315 - Tuomo Raitio, John Kane, Thomas Drugman, Christer Gobl:
HMM-based synthesis of creaky voice. 2316-2320 - Xiaoxuan Wang, Khe Chai Sim:
Integrating conditional random fields and joint multi-gram model with syllabic features for grapheme-to-phone conversion. 2321-2325 - Patrick Lehnen, Alexandre Allauzen, Thomas Lavergne, François Yvon, Stefan Hahn, Hermann Ney:
Structure learning in hidden conditional random fields for grapheme-to-phoneme conversion. 2326-2330 - Adriana Stan, Oliver Watts, Yoshitaka Mamiya, Mircea Giurgiu, Robert A. J. Clark, Junichi Yamagishi, Simon King:
TUNDRA: a multilingual corpus of found data for TTS research created with light supervision. 2331-2335 - Ranniery Maia, Mark J. F. Gales, Yannis Stylianou, Masami Akamine:
Minimum mean squared error based warped complex cepstrum analysis for statistical parametric speech synthesis. 2336-2340
ASR - Discriminative Training
- Yasser Hifny:
Augmented conditional random fields modeling based on discriminatively trained features. 2341-2344 - Karel Veselý, Arnab Ghoshal, Lukás Burget, Daniel Povey:
Sequence-discriminative training of deep neural networks. 2345-2349 - Weibin Zhang, Pascale Fung:
Discriminatively trained sparse inverse covariance matrices for low resource acoustic modeling. 2350-2354 - Yuuki Tachioka, Shinji Watanabe:
Discriminative training of acoustic models for system combination. 2355-2359 - Yan Huang, Dong Yu, Yifan Gong, Chaojun Liu:
Semi-supervised GMM and DNN acoustic model training with multi-system combination and confidence re-calibration. 2360-2364 - Jian Xue, Jinyu Li, Yifan Gong:
Restructuring of deep neural network acoustic models with singular value decomposition. 2365-2369
L2 Acquisition, Multilingualism
- Nancy F. Chen, Vivaek Shivakumar, Mahesh Harikumar, Bin Ma, Haizhou Li:
Large-scale characterization of Mandarin pronunciation errors made by native speakers of European languages. 2370-2374 - Véronique Delvaux, Kathy Huet, Myriam Piccaluga, Bernard Harmegnies:
Production training in second language acquisition: a comparison between objective measures and subjective judgments. 2375-2379 - Nicole Netelenbos, Fangfang Li:
The production and perception of voice onset time in English-speaking children enrolled in a French immersion program. 2380-2384 - Pepi Burgos, Catia Cucchiarini, Roeland van Hout, Helmer Strik:
Pronunciation errors by Spanish learners of Dutch: a data-driven study for ASR-based pronunciation training. 2385-2389 - Calbert Graham, Brechtje Post:
Realisation of tonal alignment in the English of Japanese-English late bilinguals. 2390-2394 - Agathe Benoist-Lucy, Claire Pillot-Loiseau:
The influence of language and speech task upon creaky voice use among six young American women learning French. 2395-2399
Child Computer Interaction (Special Session)
- Daniel Bone, Chi-Chun Lee, Theodora Chaspari, Matthew P. Black, Marian E. Williams, Sungbok Lee, Pat Levitt, Shrikanth S. Narayanan:
Acoustic-prosodic, turn-taking, and language cues in child-psychologist interactions for varying social demand. 2400-2404 - Hynek Boril, Qian Zhang, Pongtep Angkititrakul, John H. L. Hansen, Dongxin Xu, Jill Gilkerson, Jeffrey A. Richards:
A preliminary study of child vocalization on a parallel corpus of US and shanghainese toddlers. 2405-2409 - Felix Claus, Hamurabi Gamboa Rosales, Rico Petrick, Horst-Udo Hain, Rüdiger Hoffmann:
A survey about databases of children's speech. 2410-2414 - Vassiliki Kouloumenta, Manolis Perakakis, Alexandros Potamianos:
Affective evaluation of multimodal dialogue games for preschoolers using physiological signals. 2415-2419 - Md. Jahangir Alam, Yazid Attabi, Pierre Dumouchel, Patrick Kenny, Douglas D. O'Shaughnessy:
Amplitude modulation features for emotion recognition from speech. 2420-2424 - Daniel Bone, Chi-Chun Lee, Vikram Ramanarayanan, Shrikanth S. Narayanan, Renske S. Hoedemaker, Peter C. Gordon:
Analyzing eye-voice coordination in rapid automatized naming. 2425-2429 - Theodora Chaspari, Emily Mower Provost, Shrikanth S. Narayanan:
Analyzing the structure of parent-moderated narratives from children with ASD using an entity-based approach. 2430-2434 - Keelan Evanini, Xinhao Wang:
Automated speech scoring for non-native middle school students with multiple task types. 2435-2439 - Saeid Safavi, Peter Jancovic, Martin J. Russell, Michael J. Carey:
Identification of gender from children's speech by computers and humans. 2440-2444 - Takayuki Arai:
On why Japanese /r/ sounds are difficult for children to acquire. 2445-2449
Speaker Recognition I, II
- Achintya Kumar Sarkar, Claude Barras:
Anchor and UBM-based multi-class MLLR m-vector system for speaker verification. 2450-2454 - Leibny Paola García-Perera, Bhiksha Raj, Juan Arturo Nolazco-Flores:
Ensemble approach in speaker verification. 2455-2459 - Jun Wang, Dong Wang, Xiaojun Wu, Thomas Fang Zheng, Javier Tejedor:
Sequential model adaptation for speaker verification. 2460-2464 - Ahilan Kanagasundaram, David Dean, Javier Gonzalez-Dominguez, Sridha Sridharan, Daniel Ramos, Joaquin Gonzalez-Rodriguez:
Improving short utterance based i-vector speaker recognition using source and utterance-duration normalization techniques. 2465-2469 - Hagai Aronowitz, Oren Barkan:
On leveraging conversational data for building a text dependent speaker verification system. 2470-2473 - Wei-Qiang Zhang, Zhiyi Li, Weiwei Liu, Jia Liu:
THU-EE system fusion for the NIST 2012 speaker recognition evaluation. 2474-2478 - Daniel Garcia-Romero, Alan McCree:
Subspace-constrained supervector PLDA for speaker verification. 2479-2483 - Cong-Thanh Do, Claude Barras, Viet Bac Le, Achintya Kumar Sarkar:
Augmenting short-term cepstral features with long-term discriminative features for speaker verification of telephone data. 2484-2488 - Padmanabhan Rajan, Tomi Kinnunen, Cemal Hanilçi, Jouni Pohjalainen, Paavo Alku:
Using group delay functions from all-pole models for speaker recognition. 2489-2493 - José Portelo, Alberto Abad, Bhiksha Raj, Isabel Trancoso:
Secure binary embeddings of front-end factor analysis for privacy preserving speaker verification. 2494-2498 - Jalil Taghia, Zhanyu Ma, Arne Leijon:
On von-mises fisher mixture model in text-independent speaker identification. 2499-2503 - Mireia Díez, Amparo Varona, Mikel Peñagarikano, Luis Javier Rodríguez-Fuentes, Germán Bordel:
Using phone log-likelihood ratios as features for speaker recognition. 2504-2508 - Jesús Antonio Villalba López, Mireia Díez, Amparo Varona, Eduardo Lleida:
Handling recordings acquired simultaneously over multiple channels with PLDA. 2509-2513 - Xiao Fang, Najim Dehak, James R. Glass:
Bayesian distance metric learning on i-vector for speaker verification. 2514-2518 - Rosa González Hautamäki, Ville Hautamäki, Padmanabhan Rajan, Tomi Kinnunen:
Merging human and automatic system decisions to improve speaker recognition performance. 2519-2523
Dialog Systems and Applications I, II
- Kaisheng Yao, Geoffrey Zweig, Mei-Yuh Hwang, Yangyang Shi, Dong Yu:
Recurrent neural networks for language understanding. 2524-2528 - Korbinian Riedhammer, Van Hai Do, James Hieronymus:
A study on LVCSR and keyword search for tagalog. 2529-2533 - Sharifa Alghowinem, Roland Goecke, Michael Wagner, Julien Epps, Gordon Parker, Michael Breakspear:
Characterising depressed speech for classification. 2534-2538 - Benjamin Bigot, Grégory Senay, Georges Linarès, Corinne Fredouille, Richard Dufour:
Combining acoustic name spotting and continuous context models to improve spoken person name recognition in speech. 2539-2543 - I-Fan Chen, Chin-Hui Lee:
A resource-dependent approach to word modeling for keyword spotting. 2544-2548 - Kathryn Womack, Cecilia Ovesdotter Alm, Cara Calvelli, Jeff B. Pelz, Pengcheng Shi, Anne R. Haake:
Markers of confidence and correctness in spoken medical narratives. 2549-2553 - Ibuki Nakamura, Nobuaki Minematsu, Masayuki Suzuki, Hiroko Hirano, Chieko Nakagawa, Noriko Nakamura, Yukinori Tagawa, Keikichi Hirose, Hiroya Hashimoto:
Development of a web framework for teaching and learning Japanese prosody: OJAD (online Japanese accent dictionary). 2554-2558 - Elizabeth Shriberg, Andreas Stolcke, Suman V. Ravuri:
Addressee detection for dialog systems using temporal and spectral dimensions of speaking style. 2559-2563 - Hiroaki Hatano, Miyako Kiso, Carlos Toshinori Ishi:
Analysis of factors involved in the choice of rising or non-rising intonation in question utterances appearing in conversational speech. 2564-2568 - Asli Celikyilmaz, Gökhan Tür, Dilek Hakkani-Tür:
IsNL? a discriminative approach to detect natural language like queries for conversational understanding. 2569-2573 - Jian Cheng, Nikhil Bojja, Xin Chen:
Automatic accent quantification of indian speakers of English. 2574-2578 - Gökhan Tür, Anoop Deoras, Dilek Hakkani-Tür:
Semantic parsing using word confusion networks with conditional random fields. 2579-2583 - Sofia Strömbergsson, Anna Hjalmarsson, Jens Edlund, David House:
Timing responses to questions in dialogue. 2584-2588 - Martin Karafiát, Frantisek Grézl, Mirko Hannemann, Karel Veselý, Jan Cernocký:
BUT BABEL system for spontaneous Cantonese. 2589-2593 - Atta Norouzian, Richard C. Rose, Aren Jansen:
Semi-supervised manifold learning approaches for spoken term verification. 2594-2598 - Ying Li, Pascale Fung:
Language modeling for mixed language speech recognition using weighted phrase extraction. 2599-2603
Spoken Machine Translation and Speech Natural Language Processing I, II
- Janto Skowronek, Julian Herlinghaus, Alexander Raake:
Quality assessment of asymmetric multiparty telephone conferences: a systematic method from technical degradations to perceived impairments. 2604-2608 - Keisuke Imoto, Suehiro Shimauchi, Hisashi Uematsu, Hitoshi Ohmuro:
User activity estimation method based on probabilistic generative model of acoustic event sequence with user activity and its subordinate categories. 2609-2613 - Takatomo Kano, Shinnosuke Takamichi, Sakriani Sakti, Graham Neubig, Tomoki Toda, Satoshi Nakamura:
Generalizing continuous-space translation of paralinguistic information. 2614-2618 - Masaya Ohgushi, Graham Neubig, Sakriani Sakti, Tomoki Toda, Satoshi Nakamura:
An empirical comparison of joint optimization techniques for speech translation. 2619-2623 - Mari Ostendorf, Sangyun Hahn:
A sequential repetition model for improved disfluency detection. 2624-2628 - Henrique Medeiros, Helena Moniz, Fernando Batista, Isabel Trancoso, Luís Nunes:
Disfluency detection based on prosodic features for university lectures. 2629-2633 - Bernd T. Meyer:
What's the difference? comparing humans and machines on the Aurora 2 speech recognition task. 2634-2638 - Michele Gubian, Lou Boves, Maarten Versteegh:
Calibration of distance measures for unsupervised query-by-example. 2639-2643 - Diego Castán, Murat Akbacak:
Indexing multimedia documents with acoustic concept recognition lattices. 2644-2648 - Spyros Kousidis, Thies Pfeiffer, David Schlangen:
MINT.tools: tools and adaptors supporting acquisition, annotation and analysis of multimodal corpora. 2649-2653
Show and Tell Sessions 1-3
- Robert A. J. Clark:
Simple4all. 2654-2656 - Grégoire Pointeau, Maxime Petit, Xavier Hinaut, Guillaume Gibert, Peter Ford Dominey:
On-line learning of lexical items and grammatical constructions via speech, gaze and action-based human-robot interaction. 2657-2659 - Haruko Miyakoda:
Development of a pronunciation training system based on auditory-visual elements. 2660-2661 - Elias Azarov, Maxim Vashkevich, Denis Likhachov, Alexander A. Petrovsky:
Real-time and non-real-time voice conversion systems with web interfaces. 2662-2663 - E. Csala, Géza Németh, Csaba Zainkó:
Application of the NAO humanoid robot in the treatment of bone marrow-transplanted children (demo). 2664-2666 - Vincent Wan, Robert Anderson, Art Blokland, Norbert Braunschweiler, Langzhou Chen, BalaKrishna Kolluru, Javier Latorre, Ranniery Maia, Björn Stenger, Kayoko Yanagisawa, Yannis Stylianou, Masami Akamine, Mark J. F. Gales, Roberto Cipolla:
Photo-realistic expressive text to talking head synthesis. 2667-2669 - Ian Maddieson, Sébastien Flavier, Egidio Marsico, François Pellegrino:
Demonstration of LAPSyd: lyon-albuquerque phonological systems database. 2670-2671 - Suzanne Boyce, Marisha Speights, Keiko Ishikawa, Joel MacAuslan:
Speechmark acoustic landmark tool: application to voice pathology. 2672-2674 - Laurence Catanese, Nathan Souviraà-Labastie, Bingqing Qu, Sébastien Campion, Guillaume Gravier, Emmanuel Vincent, Frédéric Bimbot:
MODIS: an audio motif discovery software. 2675-2677
Language Model Adaptation
- Md. Akmal Haidar, Douglas D. O'Shaughnessy:
Fitting long-range information using interpolated distanced n-grams and cache models into a latent dirichlet language model for speech recognition. 2678-2682 - Yi-Wen Chen, Bo-Han Hao, Kuan-Yu Chen, Berlin Chen:
Incorporating proximity information for relevance language modeling in speech recognition. 2683-2687 - Ali Orkan Bayer, Giuseppe Riccardi:
Instance-based on-line language model adaptation. 2688-2692 - André Mansikkaniemi, Mikko Kurimo:
Unsupervised topic adaptation for morph-based speech recognition. 2693-2697 - Tim Schlippe, Lukasz Gren, Ngoc Thang Vu, Tanja Schultz:
Unsupervised language model adaptation for automatic speech recognition of broadcast news using web 2.0. 2698-2702 - Tsung-Hsien Wen, Aaron Heidel, Hung-yi Lee, Yu Tsao, Lin-Shan Lee:
Recurrent neural network based language model personalization by social network crowdsourcing. 2703-2707
Spoken Language Summarization and Understanding
- Moataz El Ayadi, Mohamed Afify:
Language-independent call routing using the large margin estimation principle. 2708-2712 - Anoop Deoras, Ruhi Sarikaya:
Deep belief network based semantic taggers for spoken language understanding. 2713-2717 - Bassam Jabaian, Fabrice Lefèvre:
Error-corrective discriminative joint decoding of automatic spoken language transcription and understanding. 2718-2722 - Catherine Lai, Jean Carletta, Steve Renals:
Detecting summarization hot spots in meetings using group level involvement and turn-taking features. 2723-2727 - Sz-Rung Shiang, Hung-yi Lee, Lin-Shan Lee:
Supervised spoken document summarization based on structured support vector machine with utterance clusters as hidden variables. 2728-2732 - Ioannis Klasinas, Alexandros Potamianos, Elias Iosif, Spiros Georgiladakis, Gianluca Mameli:
Web data harvesting for speech understanding grammar induction. 2733-2737
Speech Synthesis - Multimodal and Articulatory Synthesis
- Asterios Toutios, Shrikanth S. Narayanan:
Articulatory synthesis of French connected speech from EMA data. 2738-2742 - Xinjian Zhang, Lijuan Wang, Gang Li, Frank Seide, Frank K. Soong:
A new language independent, photo-realistic talking head driven by voice only. 2743-2747 - Chaoyang Wang, Lijuan Wang, Yasuyuki Matsushita, Bojun Huang, Magnetro Chen, Frank K. Soong:
Binocular photometric stereo acquisition and reconstruction for 3d talking head applications. 2748-2752 - Thomas Hueber, Gérard Bailly, Pierre Badin, Frédéric Elisei:
Speaker adaptation of an acoustic-articulatory inversion model using cascaded Gaussian mixture regressions. 2753-2757 - Atef Ben Youssef, Hiroshi Shimodaira, David Adam Braude:
Articulatory features for speech-driven head motion synthesis. 2758-2762 - David Adam Braude, Hiroshi Shimodaira, Atef Ben Youssef:
Template-warping based speech driven head motion synthesis. 2763-2767
Speaker Diarization and Recognition
- Anthony Larcher, Jean-François Bonastre, Benoit G. B. Fauve, Kong-Aik Lee, Christophe Lévy, Haizhou Li, John S. D. Mason, Jean-Yves Parfait:
ALIZE 3.0 - open source toolkit for state-of-the-art speaker recognition. 2768-2772 - Mohammed Senoussaoui, Patrick Kenny, Pierre Dumouchel, Najim Dehak:
New cosine similarity scorings to implement gender-independent speaker verification. 2773-2777 - Delphine Charlet, Corinne Fredouille, Géraldine Damnati, Grégory Senay:
Improving speaker identification in TV-shows using person name detection in overlaid text and speech. 2778-2782 - Mary Tai Knox, Nikki Mirghafori, Gerald Friedland:
Exploring methods of improving speaker accuracy for speaker diarization. 2783-2787 - Ryan Price, Sangeeta Biswas, Koichi Shinoda:
Combining deep speaker specific representations with GMM-SVM for speaker verification. 2788-2792 - Carola Schindler, Christoph Draxler:
Using spectral moments as a speaker specific feature in nasals and fricatives. 2793-2796
Models of Speech Perception
- Raphaël Laurent, Jean-Luc Schwartz, Pierre Bessière, Julien Diard:
A computational model of perceptuo-motor processing in speech perception: learning to imitate and categorize synthetic CV syllables. 2797-2801 - Rachel M. Theodore:
Talker-specific perceptual processing: influences on internal category structure. 2802-2806 - María Luisa García Lecumberri, Máté Attila Tóth, Yan Tang, Martin Cooke:
Elicitation and analysis of a corpus of robust noise-induced word misperceptions in Spanish. 2807-2811 - Anne Cutler, Laurence Bruggeman:
Vocabulary structure and spoken-word recognition: evidence from French reveals the source of embedding asymmetry. 2812-2816 - Odile Bagou, Ulrich H. Frauenfelder:
How do multiple sublexical cues converge in lexical segmentation? an artificial language learning study. 2817-2821 - Louis ten Bosch, Lou Boves, Mirjam Ernestus:
Towards an end-to-end computational model of speech comprehension: simulating a lexical decision task. 2822-2826
Paralinguistic Information I, II
- Sven Ewan Shepstone, Zheng-Hua Tan, Søren Holdt Jensen:
Demographic recommendation by means of group profile elicitation using speaker age and gender recognition. 2827-2831 - Nikos Malandrakis, Shiva Sundaram, Alexandros Potamianos:
Affective classification of generic audio clips using regression models. 2832-2836 - Je Hun Jeon, Duc Le, Rui Xia, Yang Liu:
A preliminary study of cross-lingual emotion recognition from speech: automatic classification versus human perception. 2837-2840 - Wenjing Han, Haifeng Li, Huabin Ruan, Lin Ma, Jiayin Sun, Björn W. Schuller:
Active learning for dimensional speech emotion recognition. 2841-2845 - Finnian Kelly, Naomi Harte:
Auditory detectability of vocal ageing and its effect on forensic automatic speaker recognition. 2846-2850 - Firoj Alam, Giuseppe Riccardi:
Comparative study of speaker personality traits recognition in conversational and broadcast news speech. 2851-2855 - Zixing Zhang, Jun Deng, Erik Marchi, Björn W. Schuller:
Active learning by label uncertainty for acoustic emotion recognition. 2856-2860 - Bo Xiao, Panayiotis G. Georgiou, Zac E. Imel, David C. Atkins, Shrikanth S. Narayanan:
Modeling therapist empathy and vocal entrainment in drug addiction counseling. 2861-2865 - Chiaki Miyazaki, Ryuichiro Higashinaka, Toshiro Makino, Yoshihiro Matsuo:
Estimating callers' levels of knowledge in call center dialogues. 2866-2870 - Juan Pablo Arias, Carlos Busso, Néstor Becerra Yoma:
Energy and F0 contour modeling with functional data analysis for emotional speech detection. 2871-2875 - Taniya Mishra, Dimitrios Dimitriadis:
Incremental emotion recognition. 2876-2880 - Cemal Hanilçi, Tomi Kinnunen, Padmanabhan Rajan, Jouni Pohjalainen, Paavo Alku, Figen Ertas:
Comparison of spectrum estimators in speaker verification: mismatch conditions induced by vocal effort. 2881-2885 - Rui Xia, Yang Liu:
Using denoising autoencoder for emotion recognition. 2886-2889
Speech and Audio Signal Processing
- Georgios Athanasopoulos, Werner Verhelst:
A phase-modified approach for TDE-based acoustic localization. 2890-2894 - Wei Xue, Shan Liang, Wenju Liu:
Interference robust DOA estimation of human speech by exploiting historical information and temporal correlation. 2895-2899 - Naomi Harte, Sadhbh Murphy, David J. Kelly, Nicola M. Marples:
Identifying new bird species from differences in birdsong. 2900-2904 - Yuri Nishigaki, Ken-Ichi Sakakibara, Masanori Morise, Ryuichi Nisimura, Toshio Irino, Hideki Kawahara:
Controlling "shout" expression in a Japanese POP singing performance: analysis and suppression study. 2905-2909 - Mahnoosh Mehrabani, John H. L. Hansen:
Dimensionality analysis of singing speech based on locality preserving projections. 2910-2914 - Md. Khademul Islam Molla, Keikichi Hirose:
Audio classification using dominant spatial patterns in time-frequency space. 2915-2919 - Tse-En Lin, Chung-Chien Hsu, Yi-Cheng Chen, Jian-Hueng Chen, Tai-Shih Chi:
Spectro-temporal modulation based singing detection combined with pitch-based grouping for singing voice separation. 2920-2923 - Jimmy Ludeña-Choez, Ascensión Gallardo-Antolín:
NMF-based temporal feature integration for acoustic event classification. 2924-2928 - Shourabh Rawat, Peter F. Schulam, Susanne Burger, Duo Ding, Yipei Wang, Florian Metze:
Robust audio-codebooks for large-scale event detection in consumer videos. 2929-2933 - M. Umair Bin Altaf, Taras Butko, Biing-Hwang Juang:
Person identification using biometric markers from footsteps sound. 2934-2938 - Wiktor Mlynarski:
Learning binaural spectrogram features for azimuthal speaker localization. 2939-2942 - Youssef Oualil, Friedrich Faubel, Dietrich Klakow:
An unsupervised Bayesian classifier for multiple speaker detection and localization. 2943-2947 - Rupayan Chakraborty, Climent Nadeu:
Joint recognition and direction-of-arrival estimation of simultaneous meeting-room acoustic events. 2948-2952 - Xiaodan Zhuang, Shuang Wu, Pradeep Natarajan, Rohit Prasad, Prem Natarajan:
Audio self organized units for high-level event detection. 2953-2957
ASR - Robustness Against Noise I-III
- Yu-Chen Kao, Berlin Chen:
Distribution-based feature normalization for robust speech recognition leveraging context and dynamics cues. 2958-2962 - Shilin Liu, Khe Chai Sim:
An investigation of temporally varying weight regression for noise robust speech recognition. 2963-2967 - Yang Li, Xunying Liu, Lan Wang:
Feature space generalized variable parameter HMMs for noise robust recognition. 2968-2972 - Philemon Brakel, Dirk Stroobandt, Benjamin Schrauwen:
Bidirectional truncated recurrent neural networks for efficient speech denoising. 2973-2977 - Ehsan Variani, Feipeng Li, Hynek Hermansky:
Multi-stream recognition of noisy speech with performance monitoring. 2978-2981 - Masakiyo Fujimoto, Tomohiro Nakatani:
Model-based noise suppression using unsupervised estimation of hidden Markov model for non-stationary noise. 2982-2986 - Karan Nathwani, Rajesh M. Hegde:
Joint noise cancellation and dereverberation using multi-channel linearly constrained minimum variance filter. 2987-2991 - Marc Delcroix, Yotaro Kubo, Tomohiro Nakatani, Atsushi Nakamura:
Is speech enhancement pre-processing still relevant when using deep neural networks for acoustic modeling? 2992-2996 - Hsin-Ju Hsieh, Berlin Chen, Jeih-weih Hung:
Histogram equalization of real and imaginary modulation spectra for noise-robust speech recognition. 2997-3001 - Bo Li, Yu Tsao, Khe Chai Sim:
An investigation of spectral restoration algorithms for deep neural networks based noise robust speech recognition. 3002-3006 - Ulpu Remes:
Bounded conditional mean imputation with an approximate posterior. 3007-3011 - Xiaodong Cui, Vaibhava Goel, Brian Kingsbury:
Mixtures of Bayesian joint factor analyzers for noise robust automatic speech recognition. 3012-3016 - Gang Liu, Dimitrios Dimitriadis, Enrico Bocchieri:
Robust speech enhancement techniques for ASR in non-stationary noise and dynamic environments. 3017-3021
Linguistic Systems, Phonetics-Phonology Interface
- Ian Maddieson, Sébastien Flavier, Egidio Marsico, Christophe Coupé, François Pellegrino:
LAPSyd: lyon-albuquerque phonological systems database. 3022-3026 - Plínio A. Barbosa:
The duration compensation issue revisited. 3027-3031 - Yoon Mi Oh, François Pellegrino, Christophe Coupé, Egidio Marsico:
Cross-language comparison of functional load for vowels, consonants, and tones. 3032-3036 - Kikuo Maekawa:
Notes on so-called inter-speaker difference in spontaneous speech: the case of Japanese voiced obstruent. 3037-3041 - Christopher Carignan, Ryan Shosted, Maojing Fu, Zhi-Pei Liang, Bradley P. Sutton:
The role of the pharynx and tongue in enhancement of vowel nasalization: a real-time MRI investigation of French nasal vowels. 3042-3046 - Margaret E. L. Renwick, Ladan Baghai-Ravary, Rosalind Temple, John S. Coleman:
Assimilation of word-final nasals to following word-initial place of articulation in UK English. 3047-3051
Speech Synthesis - Voice Conversion
- Ling-Hui Chen, Zhen-Hua Ling, Yan Song, Li-Rong Dai:
Joint spectral distribution modeling using restricted boltzmann machines for voice conversion. 3052-3056 - Zhizheng Wu, Tuomas Virtanen, Tomi Kinnunen, Engsiong Chng, Haizhou Li:
Exemplar-based unit selection for voice conversion utilizing temporal information. 3057-3061 - Hsin-Te Hwang, Yu Tsao, Hsin-Min Wang, Yih-Ru Wang, Sin-Horng Chen:
Alleviating the over-smoothing problem in GMM-based voice conversion with discriminative training. 3062-3066 - Kou Tanaka, Tomoki Toda, Graham Neubig, Sakriani Sakti, Satoshi Nakamura:
A hybrid approach to electrolaryngeal speech enhancement based on spectral subtraction and statistical voice conversion. 3067-3071 - Takuto Moriguchi, Tomoki Toda, Motoaki Sano, Hiroshi Sato, Graham Neubig, Sakriani Sakti, Satoshi Nakamura:
A digital signal processor implementation of silent/electrolaryngeal speech enhancement based on real-time statistical voice conversion. 3072-3076 - Sandesh Aryal, Daniel Felps, Ricardo Gutierrez-Osuna:
Foreign accent conversion through voice morphing. 3077-3081
Large Vocabulary Continuous Speech Recognition Systems
- Kartik Audhkhasi, Andreas M. Zavou, Panayiotis G. Georgiou, Shrikanth S. Narayanan:
Empirical link between hypothesis diversity and fusion performance in an ensemble of automatic speech recognition systems. 3082-3086 - Peter Bell, Hitoshi Yamamoto, Pawel Swietojanski, Youzheng Wu, Fergus McInnes, Chiori Hori, Steve Renals:
A lecture transcription system combining neural network acoustic and language models. 3087-3091 - Hagen Soltau, Hong-Kwang Kuo, Lidia Mangu, George Saon, Tomás Beran:
Neural network acoustic models for the DARPA RATS program. 3092-3096 - Nicola Ueffing, Maximilian Bisani, Paul Vozila:
Improved models for automatic punctuation prediction for spoken and written text. 3097-3101 - Anindya Roy, Lori Lamel, Thiago Fraga-Silva, Jean-Luc Gauvain, Ilya Oparin:
Some issues affecting the transcription of Hungarian broadcast audio. 3102-3106 - Pavel Golik, Zoltán Tüske, Ralf Schlüter, Hermann Ney:
Development of the RWTH transcription system for slovenian. 3107-3111
Robust Speaker Recognition I, II
- Naoyuki Kanda, Ryu Takeda, Yasunari Obuchi:
Noise robust speaker verification with delta cepstrum normalization. 3112-3116 - David Vandyke, Michael Wagner, Roland Goecke:
R-norm: improving inter-speaker variability modelling at the score level via regression score normalisation. 3117-3121 - Tomi Kinnunen, Md. Jahangir Alam, Pavel Matejka, Patrick Kenny, Jan Cernocký, Douglas D. O'Shaughnessy:
Frequency warping and robust speaker verification: a comparison of alternative mel-scale representations. 3122-3126 - Taufiq Hasan, John H. L. Hansen:
Acoustic factor analysis based universal background model for robust speaker verification in noise. 3127-3131 - Jesús Antonio Villalba López, Eduardo Lleida, Alfonso Ortega, Antonio Miguel:
A new Bayesian network to assess the reliability of speaker verification decisions. 3132-3136 - Weizhong Zhu, Sibel Yaman, Jason W. Pelecanos:
The IBM RATS phase II speaker recognition system: overview and analysis. 3137-3141
Acoustic and Articulatory Cues in Speech Perception
- Jason A. Shaw, Michael D. Tyler, Benjawan Kasisopa, Yuan Ma, Michael I. Proctor, Chong Han, Donald Derrick, Denis K. Burnham:
Vowel identity conditions the time course of tone recognition. 3142-3146 - Odette Scharenborg, Esther Janse:
Changes in the role of intensity as a cue for fricative categorisation. 3147-3151 - Keiichi Yasu, Takayuki Arai, Kei Kobayashi, Mitsuko Shindo:
Weighting of acoustic cues shifts to frication duration in identification of fricatives/affricates when auditory properties are degraded due to aging. 3152-3156 - Jiayin Gao, Pierre A. Hallé:
Duration as a secondary cue for perception of voicing and tone in shanghai Chinese. 3157-3161 - Marie Dekerle, Fanny Meunier, Marie-Ange N'Guyen, Estelle Gillet-Perret, Delphine Lassus-Sangosse, Sophie Donnadieu:
Development of central auditory processes and their links with language skills in typically developing children. 3162-3166 - Léo Varnet, Kenneth Knoblauch, Fanny Meunier, Michel Hoen:
Show me what you listen to! auditory classification images can reveal the processing of fine acoustic cues during speech categorization. 3167-3171
Speech Production - Data and Models
- Fabian Brackhane, Jürgen Trouvain:
The organ stop "vox humana" as a model for a vowel synthesiser. 3172-3176 - Prasanta Kumar Ghosh, Shrikanth S. Narayanan:
Information theoretic acoustic feature selection for acoustic-to-articulatory inversion. 3177-3181 - Dita Fejlová, David Lukes, Radek Skarnitzl:
Formant contours in Czech vowels: speaker-discriminating potential. 3182-3186 - Shen Liu, Jianguo Wei, Xin Wang, Wenhuan Lu, Qiang Fang, Jianwu Dang:
An anisotropic diffusion filter based on multidirectional separability. 3187-3190 - Radek Skarnitzl, Pavel Sturm, Pavel Machac:
The phonological voicing contrast in Czech: an EPG study of phonated and whispered fricatives. 3191-3195 - Shinji Maeda, Yves Laprie:
Vowel and prosodic factor dependent variations of vocal-tract length. 3196-3200 - Tijl Grootswagers, Karen Dijkstra, Louis ten Bosch, Alex Brandmeyer, Makiko Sadakata:
Word identification using phonetic features: towards a method to support multivariate fMRI speech decoding. 3201-3205 - Dhananjaya N. Gowda, Mikko Kurimo:
Analysis of breathy, modal and pressed phonation based on low frequency spectral density. 3206-3210 - Keiichi Tajima, Kuniyoshi Tanaka, Andrew Martin, Reiko Mazuka:
Is the vowel length contrast in Japanese exaggerated in infant-directed speech? 3211-3215 - Gang Chen, Robin A. Samlan, Jody Kreiman, Abeer Alwan:
Investigating the relationship between glottal area waveform shape and harmonic magnitudes through computational modeling and laryngeal high-speed videoendoscopy. 3216-3220 - Jonathan C. Kim, Hrishikesh Rao, Mark A. Clements:
Formant frequency tracking using Gaussian mixtures with maximum a posteriori adaptation. 3221-3225 - Rei Yasuda, Frank Zimmerer:
Devoicing of vowels in German, a comparison of Japanese and German speakers. 3226-3229 - Caitlin Smith, Adam C. Lammert:
Identifying consonantal tasks via measures of tongue shaping: a real-time MRI investigation of the production of vocalized syllabic /l/ in American English. 3230-3233
Speech Enhancement
- Feng Deng, Changchun Bao, Feng Bao:
A speech enhancement method by coupling speech detection and spectral amplitude estimation. 3234-3238 - Chenxi Zheng, Wai-Yip Chan:
Late reverberation suppression using MMSE modulation spectral estimation. 3239-3243 - M. A. Tugtekin Turan, Engin Erzin:
A new statistical excitation mapping for enhancement of throat microphone recordings. 3244-3248 - Nicoleta Roman, Michael I. Mandel:
Classification based binaural dereverberation. 3249-3253 - Seon Man Kim, Hong Kook Kim:
Target-to-non-target directional ratio estimation based on dual-microphone phase differences for target-directional speech enhancement. 3254-3258 - Xugang Lu, Shigeki Matsuda, Chiori Hori:
Speech spectrum restoration based on conditional restricted boltzmann machine. 3259-3263 - Faheem Khan, Ben Milner:
Speaker separation using visual speech features and single-channel audio. 3264-3268 - Wei-Lun Chuang, Kah-Meng Cheong, Chung-Chien Hsu, Tai-Shih Chi:
Spectral modulation sensitivity based perceptual acoustic echo cancellation. 3269-3273 - Vinayak Abrol, Pulkit Sharma, Anil Kumar Sao:
Speech enhancement using compressed sensing. 3274-3278 - Emad M. Grais, Hakan Erdogan:
Spectro-temporal post-enhancement using MMSE estimation in NMF based single-channel source separation. 3279-3283 - Kantapon Kaewtip, Lee Ngee Tan, Abeer Alwan:
A pitch-based spectral enhancement technique for robust speech processing. 3284-3288 - Matthew C. McCallum, Bernard J. Guillemin:
Stochastic-deterministic signal modelling for the tracking of pitch in noise and speech mixtures using factorial HMMs. 3289-3293 - Shay Maymon, Etienne Marcheret, Vaibhava Goel:
Restoration of clipped signals with application to speech recognition. 3294-3297 - Yasufumi Uezu, Keisuke Kinoshita, Mehrez Souden, Tomohiro Nakatani:
On the robustness of distributed EM based BSS in asynchronous distributed microphone array scenarios. 3298-3302
ASR - Acoustic Modeling
- Jingzhou Yang, Rogier C. van Dalen, Mark J. F. Gales:
Infinite support vector machines in speech recognition. 3303-3307 - Diego Giuliani, Fabio Brugnara:
An on-line incremental speaker adaptation technique for audio stream transcription. 3308-3312 - Dominic Telaar, Mark C. Fuhs:
Accent- and speaker-specific polyphone decision trees for non-native speech recognition. 3313-3316 - Simon Wiesler, Jinyu Li, Jian Xue:
Investigations on hessian-free optimization for cross-entropy training of deep neural networks. 3317-3321 - Masahiro Saiko, Shigeki Matsuda, Ken Hanazawa, Ryosuke Isotani, Chiori Hori:
Cross-lingual acoustic model adaptation based on transfer vector field smoothing with MAP. 3322-3326 - Hiroshi Fujimura, Yusuke Shinohara, Takashi Masuko:
N-best rescoring by phoneme classifiers using subclass adaboost algorithm. 3327-3331 - Tetsuji Ogawa, Feipeng Li, Hynek Hermansky:
Stream selection and integration in multistream ASR using GMM-based performance monitoring. 3332-3336 - Néstor Becerra Yoma, Claudio Garretón, Fernando Huenupán, Ignacio Catalan, Jorge Wuth:
VTLN based on the linear interpolation of contiguous mel filter-bank energies. 3337-3341 - Fabian Triefenbach, Azarakhsh Jalalvand, Kris Demuynck, Jean-Pierre Martens:
Context-dependent modeling and speaker normalization applied to reservoir-based phone recognition. 3342-3346 - Thiago Fraga-Silva, Jean-Luc Gauvain, Lori Lamel:
Interpolation of acoustic models for speech recognition. 3347-3351 - Muhammad Ali Tahir, Heyun Huang, Ralf Schlüter, Hermann Ney, Louis ten Bosch, Bert Cranen, Lou Boves:
Training log-linear acoustic models in higher-order polynomial feature space for speech recognition. 3352-3355 - Venkata Neelima Parinam, Chandra Sekhar Vootkuri, Stephen A. Zahorian:
Comparison of spectral analysis methods for automatic speech recognition. 3356-3360 - D. Rama Sanand, Torbjørn Svendsen:
Synthetic speaker models using VTLN to improve the performance of children in mismatched speaker conditions for ASR. 3361-3365 - Ossama Abdel-Hamid, Li Deng, Dong Yu:
Exploring convolutional neural network structures and optimization techniques for speech recognition. 3366-3370
Special Event: ESCA/ISCA Anniversary
- Joseph Mariani, Patrick Paroubek, Gil Francopoulo, Marine Delaborde:
Rediscovering 25 years of discoveries in spoken language processing: a preliminary ISCA archive analysis. 3371-3403
Language Modeling for Conversational Speech
- M. Ali Basha Shaik, Amr El-Desoky Mousa, Ralf Schlüter, Hermann Ney:
Feature-rich sub-lexical language models using a maximum entropy approach for German LVCSR. 3404-3408 - Amr El-Desoky Mousa, M. Ali Basha Shaik, Ralf Schlüter, Hermann Ney:
Morpheme level hierarchical pitman-yor class-based language models for LVCSR of morphologically rich languages. 3409-3413 - Benjamin Lambert, Bhiksha Raj, Rita Singh:
Discriminatively trained dependency language modeling for conversational speech recognition. 3414-3418 - Yujing Si, Qingqing Zhang, Ta Li, Jielin Pan, Yonghong Yan:
Prefix tree based n-best list re-scoring for recurrent neural network language model used in speech recognition system. 3419-3423 - Xunying Liu, Mark J. F. Gales, Philip C. Woodland:
Cross-domain paraphrasing for improving language modelling using out-of-domain data. 3424-3428 - Ryo Masumura, Hirokazu Masataki, Takanobu Oba, Osamu Yoshioka, Satoshi Takahashi:
Viterbi decoding for latent words language models using gibbs sampling. 3429-3433
Speech Enhancement and Coding
- Tom Bäckström:
Computationally efficient objective function for algebraic codebook optimization in ACELP. 3434-3438 - Sebastian Möller, Emilia Kelaidi, Friedemann Köster, Nicolas Côté, Patrick Bauer, Tim Fingscheidt, Thomas Schlien, Hannu Pulakka, Paavo Alku:
Speech quality prediction for artificial bandwidth extension algorithms. 3439-3443 - Bingyin Xia, Changchun Bao:
Speech enhancement with weighted denoising auto-encoder. 3444-3448 - Milos Cernak, Xingyu Na, Philip N. Garner:
Syllable-based pitch encoding for low bit rate speech coding with recognition/synthesis architecture. 3449-3452 - Nguyen Duc Duy, Masayuki Suzuki, Nobuaki Minematsu, Keikichi Hirose:
Artificial bandwidth extension based on regularized piecewise linear mapping with discriminative region weighting and long-Span features. 3453-3457 - Bong-Ki Lee, Chungsoo Lim, Jihwan Park, Joon-Hyuk Chang:
Enhanced muting method in packet loss concealment of ITU-t g.722 employing optimized sigmoid function. 3458-3462
Spoken Machine Translation and Speech Natural Language Processing I, II
- Benoît Favre, Kyla Cheung, Siavash Kazemian, Adam Lee, Yang Liu, Cosmin Munteanu, Ani Nenkova, Dennis Ochei, Gerald Penn, Stephen Tratz, Clare R. Voss, Frauke Zeller:
Automatic human utility evaluation of ASR systems: does WER really predict performance? 3463-3467 - Vivek Kumar Rangarajan Sridhar, John Chen, Srinivas Bangalore:
Corpus analysis of simultaneous interpretation data for improving real time speech translation. 3468-3472 - Eunah Cho, Christian Fügen, Teresa Herrmann, Kevin Kilgour, Mohammed Mediani, Christian Mohr, Jan Niehues, Kay Rottmann, Christian Saam, Sebastian Stüker, Alex Waibel:
A real-world system for simultaneous translation of German lectures. 3473-3477 - Dekai Wu, Karteek Addanki, Markus Saers:
Freestyle: a challenge-response system for hip hop lyrics via unsupervised induction of stochastic transduction grammars. 3478-3482 - Andreas Tsiartas, Panayiotis G. Georgiou, Shrikanth S. Narayanan:
Toward transfer of acoustic cues of emphasis across languages. 3483-3486 - Tomoki Fujita, Graham Neubig, Sakriani Sakti, Tomoki Toda, Satoshi Nakamura:
Simple, lexicalized choice of translation timing for simultaneous speech translation. 3487-3491
ASR - Robustness Against Noise I-III
- Liang Lu, Arnab Ghoshal, Steve Renals:
Noise adaptive training for subspace Gaussian mixture models. 3492-3496 - George Saon, Samuel Thomas, Hagen Soltau, Sriram Ganapathy, Brian Kingsbury:
The IBM speech activity detection system for the DARPA RATS program. 3497-3501 - Armin Sehr, Takuya Yoshioka, Marc Delcroix, Keisuke Kinoshita, Tomohiro Nakatani, Roland Maas, Walter Kellermann:
Conditional emission densities for combining speech enhancement and recognition systems. 3502-3506 - Martin Wolf, Climent Nadeu:
Channel selection using n-best hypothesis for multi-microphone ASR. 3507-3511 - Takaaki Ishii, Hiroki Komiyama, Takahiro Shinozaki, Yasuo Horiuchi, Shingo Kuroiwa:
Reverberant speech recognition based on denoising autoencoder. 3512-3516 - Shay Maymon, Pierre L. Dognin, Xiaodong Cui, Vaibhava Goel:
Adaptive stereo-based stochastic mapping. 3517-3521
Articulatory and Acoustic Cues of Speech Prosody
- Thi Lan Nguyen, Alexis Michaud, Do Dat Tran, Dang-Khoa Mac:
The interplay of intonation and complex lexical tones: how speaker attitudes affect the realization of glottalization on vietnamese sentence-final particles. 3522-3526 - Ailbhe Ní Chasaide, Irena Yanushevskaya, John Kane, Christer Gobl:
The voice prominence hypothesis: the interplay of F0 and voice source features in accentuation. 3527-3531 - Albert Lee, Yi Xu, Santitham Prom-on:
Mora-based pre-low raising in Japanese pitch accent. 3532-3536 - Hélène Loevenbruck, Mohamed Ameur Ben Jannet, Mariapaola D'Imperio, Mathilde Spini, Maud Champagne-Lavau:
Prosodic cues of sarcastic speech in French: slower, higher, wider. 3537-3541 - Lucie Ménard, Annie Leclerc, Mark K. Tiede, Amélie Prémont, Christine Turgeon, Paméla Trudeau-Fisette, Dominique Côté:
Correlates of contrastive focus in congenitally blind adults and sighted adults. 3542-3546 - Laurianne Georgeton, Nicolas Audibert:
Is protrusion of French rounded vowels affected by prosodic positions? 3547-3551
Intelligibility-Enhancing Speech Modifications (Special Session)
- Martin Cooke, Catherine Mayo, Cassia Valentini-Botinhao:
Intelligibility-enhancing speech modifications: the hurricane challenge. 3552-3556 - Daniel Erro, Tudor-Catalin Zorila, Yannis Stylianou, Eva Navas, Inma Hernáez:
Statistical synthesizer with embedded prosodic and spectral modifications to generate highly intelligible speech in noise. 3557-3561 - Antti Suni, Reima Karhila, Tuomo Raitio, Mikko Kurimo, Martti Vainio, Paavo Alku:
Lombard modified text-to-speech synthesis for improved intelligibility: submission for the hurricane challenge 2013. 3562-3566 - Cassia Valentini-Botinhao, Junichi Yamagishi, Simon King, Yannis Stylianou:
Combining perceptually-motivated spectral shaping with loudness and duration modification for intelligibility enhancement of HMM-based synthetic speech in noise. 3567-3571 - Elizabeth Godoy, Yannis Stylianou:
Increasing speech intelligibility via spectral shaping with frequency warping and dynamic range compression plus transient enhancement. 3572-3576 - Henning F. Schepker, Jan Rennies, Simon Doclo:
Improving speech intelligibility in noise by SII-dependent preprocessing using frequency-dependent amplification and dynamic range compression. 3577-3581 - Cees H. Taal, Jesper Jensen:
SII-based speech preprocessing for intelligibility improvement in noise. 3582-3586 - Mengqiu Zhang, Petko Nikolov Petkov, W. Bastiaan Kleijn:
Rephrasing-based speech intelligibility enhancement. 3587-3591 - Vincent Aubanel, Martin Cooke:
Information-preserving temporal reallocation of speech in the presence of fluctuating maskers. 3592-3596 - Petko Nikolov Petkov, W. Bastiaan Kleijn:
Preservation of speech spectral dynamics enhances intelligibility. 3597-3601 - Henk Brouckxon, Werner Verhelst:
An overview of the VUB entry for the 2013 hurricane challenge. 3602-3604 - Reiko Takou, Nobumasa Seiyama, Atsushi Imai:
Improvement of speech intelligibility by reallocation of spectral energy. 3605-3607
Speech Technology for Speech and Hearing Disorders I, II
- Bart Vaerenberg, Louis ten Bosch, Wojtek Kowalczyk, Martine Coene, Herwig De Smet, Paul J. Govaerts:
Language-universal speech audiometry with automated scoring. 3608-3612 - Annemiek Hammer, Bart Vaerenberg, Wojtek Kowalczyk, Louis ten Bosch, Martine Coene, Paul J. Govaerts:
Balancing word lists in speech audiometry through large spoken language corpora. 3613-3616 - Verónica López-Ludeña, Rubén San Segundo, Javier Ferreiros, José M. Pardo, E. Ferreiro:
Developing an information system for deaf. 3617-3621 - Myung Jong Kim, Joohong Yoo, Hoirin Kim:
Dysarthric speech recognition using dysarthria-severity-dependent and speaker-adaptive models. 3622-3626 - Ghulam Muhammad, Moutasem Melhem:
Voice pathology detection and classification using MPEG-7 audio low-level features. 3627-3631 - Abdellah Kacha, Francis Grenez, Jean Schoentgen:
Empirical mode decomposition-based spectral acoustic cues for disordered voices analysis. 3632-3636 - Ryo Aihara, Ryoichi Takashima, Tetsuya Takiguchi, Yasuo Ariki:
Exemplar-based individuality-preserving voice conversion for articulation disorders in noisy environments. 3637-3641 - Heidi Christensen, Magda B. Aniol, Peter Bell, Phil D. Green, Thomas Hain, Simon King, Pawel Swietojanski:
Combining in-domain and out-of-domain speech data for automatic recognition of disordered speech. 3642-3645 - Guangting Mai, James W. Minett, William S.-Y. Wang:
Effects of envelope filter cutoff frequency on the intelligibility of Mandarin noise-vocoded speech in babble noise: implications for cochlear implants. 3646-3650
Robust Speaker Recognition I, II
- Kong-Aik Lee, Anthony Larcher, Chang Huai You, Bin Ma, Haizhou Li:
Multi-session PLDA scoring of i-vector for partially open-set speaker detection. 3651-3655 - Keith W. Godin, Seyed Omid Sadjadi, John H. L. Hansen:
Impact of noise reduction and spectrum estimation on noise robust speaker identification. 3656-3660 - Takanori Yamada, Longbiao Wang, Atsuhiko Kai:
Improvement of distant-talking speaker identification using bottleneck features of DNN. 3661-3664 - Alessio Brutti, Maurizio Omologo:
Geometric contamination for GMM/UBM speaker verification in reverberant environments. 3665-3669 - Richard D. McClanahan, Phillip L. De Leon:
Towards a more efficient SVM supervector speaker verification system using Gaussian reduction and a tree-structured hash. 3670-3673 - Ahilan Kanagasundaram, David Dean, Javier Gonzalez-Dominguez, Sridha Sridharan, Daniel Ramos, Joaquin Gonzalez-Rodriguez:
Improving the PLDA based speaker verification in limited microphone data conditions. 3674-3678 - Jesús Antonio Villalba López, Eduardo Lleida, Alfonso Ortega, Antonio Miguel:
The I3a speaker recognition system for NIST SRE12: post-evaluation analysis. 3679-3683 - Themos Stafylakis, Patrick Kenny, Pierre Ouellet, Javier Perez, Marcel Kockmann, Pierre Dumouchel:
Text-dependent speaker recognition using PLDA with uncertainty propagation. 3684-3688 - Sri Harish Reddy Mallidi, Sriram Ganapathy, Hynek Hermansky:
Robust speaker recognition using spectro-temporal autoregressive models. 3689-3693 - Padmanabhan Rajan, Tomi Kinnunen, Ville Hautamäki:
Effect of multicondition training on i-vector PLDA configurations for speaker recognition. 3694-3697 - Mitchell McLaren, Victor Abrash, Martin Graciarena, Yun Lei, Jan Pesán:
Improving robustness to compressed speech in speaker recognition. 3698-3702 - Vikramjit Mitra, Mitchell McLaren, Horacio Franco, Martin Graciarena, Nicolas Scheffer:
Modulation features for noise robust speaker identification. 3703-3707 - Ville Hautamäki, You-Chi Cheng, Padmanabhan Rajan, Chin-Hui Lee:
Minimax i-vector extractor for short duration speaker verification. 3708-3712 - Mike Fowler, Mark McCurry, Jonathan Bramsen, Kehinde Dunsin, Jeremiah Remus:
Standoff speaker recognition: effects of recording distance mismatch on speaker recognition system performance. 3713-3716
Dialog Systems and Applications I, II
- Sofia Strömbergsson, Christina Tånnander:
Correlates to intelligibility in deviant child speech - comparing clinical evaluations to audience response system-based evaluations by untrained listeners. 3717-3721 - Kathryn Womack, Cecilia Ovesdotter Alm, Cara Calvelli, Jeff B. Pelz, Pengcheng Shi, Anne R. Haake:
Using linguistic analysis to characterize conceptual units of thought in spoken medical narratives. 3722-3726 - Francesco Cutugno, Alberto Finzi, Michelangelo Fiore, Enrico Leone, Silvia Rossi:
Interacting with robots via speech and gestures, an integrated architecture. 3727-3731 - Mohamed Hatmi, Christine Jacquin, Emmanuel Morin, Sylvain Meignier:
Incorporating named entity recognition into the speech transcription process. 3732-3736 - Teppei Ohno, Tomoyosi Akiba:
DTW-distance-ordered spoken term detection. 3737-3741 - Sangkeun Jung, Seung-Hoon Na:
Refining sentence similarity with discourse information in dialog system. 3742-3746 - Ryohei Nakatani, Tetsuya Takiguchi, Yasuo Ariki:
Two-step correction of speech recognition errors based on n-gram and long contextual information. 3747-3750 - Sumit Negi, Ramnath Balasubramanyan, Santanu Chaudhury:
Inferring actor communities from videos. 3751-3755 - Xavier Bost, Marc El-Bèze, Renato De Mori:
Multiple topic identification in telephone conversations. 3756-3760 - Wei Chen, Sankaranarayanan Ananthakrishnan, Rohit Prasad, Prem Natarajan:
Variable-Span out-of-vocabulary named entity detection. 3761-3765 - Andrew L. Kun, Oskar Palinko, Zeljko Medenica, Peter A. Heeman:
On the feasibility of using pupil diameter to estimate cognitive load changes for in-vehicle spoken dialogues. 3766-3770 - Grégoire Mesnil, Xiaodong He, Li Deng, Yoshua Bengio:
Investigation of recurrent-neural-network architectures and learning methods for spoken language understanding. 3771-3775 - Xiaohu Liu, Ruhi Sarikaya, Chris Brockett, Chris Quirk, William B. Dolan:
Paraphrase features to improve natural language understanding. 3776-3779 - Dilek Hakkani-Tür, Asli Celikyilmaz, Larry P. Heck, Gökhan Tür:
A weakly-supervised approach for discovering new user intents from search query logs. 3780-3784 - Puyang Xu, Ruhi Sarikaya:
Exploiting shared information for multi-intent natural language sentence classification. 3785-3789
Special Event: ESCA/ISCA Anniversary
- Hiroya Fujisaki:
An inter- and cross-disciplinary perspective of spoken language processing. - Roger K. Moore:
Progress and prospects for speech technology: what ordinary people think.
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.