default search action
ICSLP 1992: Banff, Alberta, Canada
- The Second International Conference on Spoken Language Processing, ICSLP 1992, Banff, Alberta, Canada, October 13-16, 1992. ISCA 1992
- Peter Ladefoged:
Knowing enough to analyze spoken languages. 1-4 - Renato de Mori, Roland Kuhn:
Speech understanding strategies based on string classification trees. 441-448 - Patricia K. Kuhl:
Infants' perception and representation of speech: development of a new theory. 449-456 - Hajime Hirose:
The behavior of the larynx in spoken language production. 457-458
Word Spotting in ASR
- Eduardo Lleida, José B. Mariño, Josep M. Salavedra, Antonio Bonafonte:
Syllabic fillers for Spanish HMM keyword spotting. 5-8 - Yasuhiro Komori, David Rain Ton:
Minimum error classification training for HMM-based keyword spotting. 9-12 - Gregory J. Clary, John H. L. Hansen:
A novel speech recognizer for keyword spotting. 13-16 - Herbert Gish, Kenney Ng, Jan Robin Rohlicek:
Secondary processing using speech segments for an HMM word spotting system. 17-20 - Ming-Whei Feng, Baruch Mazor:
Continuous word spotting for applications in telecommunications. 21-24 - Maurizio Copperi:
A low bit-rate CELP coder based on multi-path search methods. 25-28
Speech Coding
- Katsushi Seza, Hirohisa Tasaki, Shinya Takahashi:
Fully vector quantized arm a analysis combined with glottal model for low bit rate coding. 29-32 - Erdal Paksoy, Wai-Yip Chan, Allen Gersho:
Vector quantization of speech LSF parameters with generalized product codes. 33-36 - Yair Shoham:
Low-rate speech coding based on time-frequency interpolation. 37-40 - Tomohiko Taniguchi, Yoshinori Tanaka, Yasuji Ohta, Fumio Amano:
Improved CELP speech coding at 4 kbit/s and below. 41-44
Speech Production: Coarticulation
- Antonio Bonafonte, José B. Mariño, Montse Pardàs:
Efficient integration of coarticulation and lexical information in a finite state grammar. 45-48 - Herbert A. Leeper, A. P. Rochet, Ian R. A. MacKay:
Characteristics of nasalance in canadian speakers of English and French. 49-52 - Christine H. Shadle, Andre Moulinier, Christian U. Dobelke, Celia Scully:
Ensemble averaging applied to the analysis of fricative consonants. 53-56 - Andrew Slater, Sarah Hawkins:
Effects of stress and vowel context on velar stops in british English. 57-60 - Emanuela Magno Caldognetto, Kyriaki Vagges, Giancarlo Ferrigno, Maria Grazia Busà:
Lip rounding coarticulation in Italian. 61-64
Auditory Models
- Paula M. T. Smeele, Anne C. Sittig, Vincent J. van Heuven:
Intelligibility of audio-visually desynchronised speech: asymmetrical effect of phoneme position. 65-68 - Unto K. Laine:
Speech analysis using complex orthogonal auditory transform (coat). 69-72 - Yuqing Gao, Taiyi Huang, Shaoyan Chen, Jean Paul Haton:
Auditory model based speech processing. 73-76 - Gary N. Tajchman, Nathan Intrator:
Phonetic classification of timit segments preprocessed with lyon's cochlear model using a supervised/unsupervised hybrid neural network. 77-80 - Thomas Holton, Steve Love, Stephen P. Gill:
Formant and pitch-pulse detection using models of auditory signal processing. 81-84
Recognition of Telephone Speech
- Hynek Hermansky, Nelson Morgan:
Towards handling the acoustic environment in spoken language processing. 85-88 - Alberto Ciaramella, Davide Clementino, Roberto Pacifici:
Real-time speaker-independent large-vocabulary CDHMM-based continuous telephonic speech recognizer. 89-92 - Matthew Lennig, Douglas Sharp, Patrick Kenny, Vishwa Gupta, Kristin Precoda:
Flexible vocabulary recognition of speech. 93-96 - Benjamin Chigier, Hong C. Leung:
The effects of signal representations, phonetic classification techniques, and the telephone network. 97-100
Text-to-Speech Synthesis 1, 1
- Leon Gulikers, Rijk Willemse:
A lexicon for a text-to-speech system. 101-104 - Rijk Willemse, Leon Gulikers:
Word class assignment in a text-to-speech system. 105-108 - Gösta Bruce, Björn Granström, Kjell Gustafson, David House:
Aspects of prosodic phrasing in Swedish. 109-112 - Kirk P. H. Sullivan, Robert I. Damper:
Synthesis-by-analogy: a bilingual investigation using German and English. 113-116 - Leonard C. Manzara, David R. Hill:
Degas: a system for rule-based diphone speech synthesis. 117-120 - Shyam Sunder Agrawal, Kenneth N. Stevens:
Towards synthesis of Hindi consonants using KLSYN88. 177-180 - Louis C. W. Pols:
Multi-lingual synthesis evaluation methods. 181-184 - Björn Granström, Petur Helgason, Hoskuldur Thrainsson:
The interaction of phonetics, phonology and morphology in an icelandic text-to-speech system. 185-188
Voice Source Characteristics
- Helmer Strik, Joop Jansen, Louis Boves:
Comparing methods for automatic extraction of voice source parameters from continuous speech. 121-124 - Jacques C. Koreman, Louis Boves, Bert Cranen:
The influence of linguistic variations on the voice source characteristics. 125-128 - Sarah K. Palmer, Jill House:
Dynamic voice source changes in natural and synthetic speech. 129-132 - Satoshi Imaizumi, Jan Gauffin:
Acoustic and perceptual modelling of the voice quality caused by fundamental frequency perturbation. 133-136 - Shigeru Kiritani, Hiroshi Imagawa, Hajime Hirose:
Vocal cord vibration during consonants - high-speed digital imaging using a fiberscope. 1661-1664
Speech Perception: Higher-Order Processes 1, 1
- David R. Traum, James F. Allen:
A "speech acts" approach to grounding in conversation. 137-140 - Sheila Meltzer:
Antecedent activation by empty pronominals in Spanish. 141-144 - Ron Smyth:
Multiple feature matching in pronoun resolution: a new look at parallel function. 145-148 - Keh-Yih Su, Jing-Shin Chang, Yi-Chung Lin:
A discriminative approach for ambiguity resolution based on a semantic score function. 149-152 - Nobuaki Minematsu, Sumio Ohno, Keikichi Hirose, Hiroya Fujisaki:
The influence of semantic and syntactic information on spoken sentence recognition. 153-156 - Lynne C. Nygaard, Mitchell Sommers, David B. Pisoni:
Effects of speaking rate and talker variability on the representation of spoken words in memory. 209-212 - Hugo Quené, Yvette Smits:
On the absence of word segmentation at "weak" syllables. 213-216 - Mitchell Sommers, Lynne C. Nygaard, David B. Pisoni:
Stimulus variability and the perception of spoken words: effects of variations in speaking rate and overall amplitude. 217-220 - James M. McQueen, Anne Cutler:
Words within words: lexical statistics and lexical access. 221-224
Speaker-Independent Word Recognition
- Stephan Euler, Joachim Zinke:
Experiments on the use of the generalized probabilistic descent method in speech recognition. 157-160 - Ricardo de Córdoba, José Manuel Pardo, José Colás:
Improving and optimizing speaker independent, 1000 words speech recognition in Spanish. 161-164 - John F. Pitrelli, David M. Lubensky, Benjamin Chigier, Hong C. Leung:
Multiple-level evaluation of speech recognition systems. 165-168 - Tatsuya Kimura, Mitsuru Endo, Shoji Hiraoka, Katsuyuki Niyada:
Speaker independent word recognition using continuous matching of parameters in time-spectral form based on statistical measure. 169-172 - R. Roddeman, H. Drexler, Louis Boves:
Automatic derivation of lexical models for a very large vocabulary speech recognition system. 173-176
Human Factors
- Anne Cutler, Tony Robinson:
Response time as a metric for comparison of speech recognition by humans and machines. 189-192 - S. M. (Raj) Ulagaraj:
Characterization of directory assistance operator-customer dialogues in AGT limited. 193-196 - Sheri Hunnicutt, Lynette Hirschman, Joseph Polifroni, Stephanie Seneff:
Analysis of the effectiveness of system error messages in a human-machine travel planning task. 197-200 - David Goodine, Lynette Hirschman, Joseph Polifroni, Stephanie Seneff, Victor Zue:
Evaluating interactive spoken language systems. 201-204 - Ute Jekosch:
The cluster-identification test. 205-208
Continuous Speech Recognition 1, 2
- Patrick Kenny, Rene Hollan, Gilles Boulianne, Harinath Garudadri, Yan Ming Cheng, Matthew Lennig, Douglas D. O'Shaughnessy:
Experiments in continuous speech recognition with a 60, 000 word vocabulary. 225-228 - Gilles Boulianne, Patrick Kenny, Matthew Lennig, Douglas D. O'Shaughnessy, Paul Mermelstein:
HMM training on unconstrained speech for large vocabulary, continuous speech recognition. 229-232 - David Rainion, Shigeki Sagayama:
Appropriate error criterion selection for continuous speech HMM minimum error training. 233-236 - Akito Nagai, Kenji Kita, Toshiyuki Hanazawa, Tadashi Suzuki, Tomohiro Iwasaki, Tsuyoshi Kawabata, Kunio Nakajima, Kiyohiro Shikano, Tsuyoshi Morimoto, Shigeki Sagayama, Akira Kurematsu:
Hardware implementation of realtime 1000-word HMM-LR continuous speech recognition. 237-240 - Madeleine Bates, Robert J. Bobrow, Pascale Fung, Robert Ingria, Francis Kubala, John Makhoul, Long Nguyen, Richard M. Schwartz, David Stallard:
Design and performance of HARC, the BBN spoken language understanding system. 241-244 - Otoya Shirotsuka, G. Kawai, Michael Cohen, Jared Bernstein:
Performance of speaker-independent Japanese recognizer as a function of training set size and diversity. 297-300 - Kouichi Yamaguchi, Shigeki Sagayama, Kenji Kita, Frank K. Soong:
Continuous mixture HMM-LR using the a* algorithm for continuous speech recognition. 301-304 - Kenji Kita, Tsuyoshi Morimoto, Kazumi Ohkura, Shigeki Sagayama:
Continuously spoken sentence recognition by HMM-LR. 305-308 - Akinori Ito, Shozo Makino:
Word pre-selection using a redundant hash addressing method for continuous speech recognition. 309-312 - Andrej Ljolje, Michael D. Riley:
Optimal speech recognition using phone recognition and lexical access. 313-316
Natural Language Processing and Speech Understanding 1-3
- Nick Waegner, Steve J. Young:
A trellis-based language model for speech recognition. 245-248 - Carla B. Zoltowski, Mary P. Harper, Leah H. Jamieson, Randall A. Helzerman:
PARSEC: a constraint-based framework for spoken language understanding. 249-252 - Gareth J. F. Jones, Jeremy H. Wright, E. N. Wrigley:
The HMM interface with hybrid grammar-bigram language models for speech recognition. 253-256 - Atsuhiko Kai, Seiichi Nakagawa:
A frame-synchronous continuous speech recognition algorithm using a top-down parsing of context-free grammar. 257-260 - Fernando Pereira, David B. Roe:
Empirical properties of finite state approximations for phrase structure grammars. 261-264 - Stephanie Seneff, Helen M. Meng, Victor Zue:
Language modelling for recognition and understanding using layered bigrams. 317-320 - David Goddeau:
Using probabilistic shift-reduce parsing in speech recognition systems. 321-324 - Tim Howells, David Friedman, Mark A. Fanty:
Broca, an integrated parser for spoken language. 325-328 - P. V. S. Rao, Nandini Bondale:
Blank slate language processor for speech recognition. 329-332 - Eric Jackson:
Integrating two complementary approaches to spoken language understanding. 333-336 - Marcello Pelillo, Mario Refice:
Learning compatibility coefficients for word-class disambiguation relaxation processes. 389-392 - Kaichiro Hatazaki, Jun Noguchi, Akitoshi Okumura, Kazunaga Yoshida, Takao Watanabe:
INTERTALKER: an experimental automatic interpretation system using conceptual representation. 393-396 - Tsuyoshi Morimoto, Toshiyuki Takezawa, Kazumi Ohkura, Masaaki Nagata, Fumihiro Yato, Shigeki Sagayama, Akira Kurematsu:
Enhancement of ATR's spoken language translation system: SL-TRANS2. 397-400 - Tsuyoshi Morimoto:
Continuous speech recognition using a combination of syntactic constraints and dependency relationship. 401-404 - Roberto Pieraccini, Zakhar Gorelov, Esther Levin, Evelyne Tzoukermann:
Automatic learning in spoken language understanding. 405-408
Language Learning and Acquisition 1, 2
- Michael P. Robb, Harold R. Bauer:
Prespeech and early speech coarticulation: american English and Japanese characteristics. 265-268 - Hiroaki Kojima, Kazuyo Tanaka, Satoru Hayamizu:
Formation of phonological concept structures from spoken word samples. 269-272 - Bernard L. Rochet, Fangxin Chen:
Acquisition of the French VOT contrasts by adult speakers of Mandarin Chinese. 273-276 - Michael Gasser:
Phonology as a byproduct of learning to recognize and produce words: a connectionist model. 277-280 - Michael S. Hurlburt, Judith C. Goodman:
The development of lexical effects on children's phoneme identifications. 337-340 - Pierre A. Hallé, Benedicte de Boysson-Bardies:
Word recognition before production of first words? 341-344 - Toshisada Deguchi, Shigeru Kiritani, Akiko Hayashi, Fumi Katoh:
The effect of fundamental frequency for vowel perception in infants. 345-348
Speech Perception: Units of Processing
- William C. Treurniet:
Objective measurement of phoneme similarity. 281-284 - Michael D. Riley, Andrej Ljolje:
Recognizing phonemes vs. recognizing phones: a comparison. 285-288 - Bruce L. Derwing, Terrance M. Nearey, R. A. Beinert, T. A. Bendrien:
On the role of the segment in speech processing by human listeners: evidence from speech perception and from global sound similarity judgments. 289-292 - Grace E. Wiebe, Bruce L. Derwing:
The syllabic status of postvocalic resonants in an unwritten low German dialect. 293-296
Prosody: The Phrase and Beyond 1, 2
- Agaath M. C. Sluijter, Vincent J. van Heuven, Anneke Neijt:
The influence of focus distribution and lexical stress on the temporal organisation of the syllable. 349-352 - Agaath M. C. Sluijter, Jacques M. B. Terken:
The development and perceptive evaluation of a model for paragraph intonation in dutch. 353-356 - Nobuyoshi Kaiki, Yoshinori Sagisaka:
Pause characteristics and local phrase-dependency structure in Japanese. 357-360 - Bernd Möbius, Matthias Pätzold:
F0 synthesis based on a quantitative model of German intonation. 361-364 - Kenneth N. Ross, Mari Ostendorf, Stefanie Shattuck-Hufnagel:
Factors affecting pitch accent placement. 365-368 - Marc Swerts, Ronald Geluykens, Jacques M. B. Terken:
Prosodic correlates of discourse units in spontaneous speech. 421-424 - Shin'ya Nakajima, James F. Allen:
Prosody as a cue for discourse structure. 425-428 - Barbara J. Grosz, Julia Hirschberg:
Some intonational characteristics of discourse structure. 429-432 - Hiroya Fujisaki, Keikichi Hirose, Haitao Lei:
Prosody and syntax in spoken sentences of standard Chinese. 433-436 - Kathleen Bishop:
Modeling sentential stress in the context of a large vocabulary continuous speech recognizer. 437-440
Speaker Adaptation
- Kazumi Ohkura, Masahide Sugiyama, Shigeki Sagayama:
Speaker adaptation based on transfer vector field smoothing with continuous mixture density HMMs. 369-372 - Tatsuo Matsuoka, Kiyohiro Shikano:
Speaker adaptation by modifying mixture coefficients of speaker-independent mixture Gaussian HMMs. 373-376 - Yifan Gong, Olivier Siohan, Jean Paul Haton:
Minimization of speech alignment error by iterative transformation for speaker adaptation. 377-380 - Hiroaki Hattori, Shigeki Sagayama:
Vector field smoothing principle for speaker adaptation. 381-384 - Tetsunori Kobayashi, Katsuhiko Shirai:
Spectral mapping onto probabilistic domain using neural networks and its application to speaker adaptive phoneme recognition. 385-388
Pronunciation Training
- Jean-Paul Lefèvre, Mervyn A. Jack, Claudio Maggio, Mario Refice, Fabio Gabrieli, Michelina Savino, Luigi Santangelo:
An interactive system for automated pronunciation improvement. 409-412 - Edmund Rooney, Steven M. Hiller, John Laver, Mervyn A. Jack:
Prosodic features for automated pronunciation improvement in the spell system. 413-416 - Maria-Gabriella Di Benedetto, Fabrizio Carraro, Steven M. Hiller, Edmund Rooney:
Vowels pronunciation assessment in the spell system. 417-420
Self-Organizing Systems in ASR 1, 2
- Franck Poirier:
Self-organizing map with supervision for speech recognition. 459-462 - Gregory R. De Haan, Ömer Egecioglu:
Topology preservation for speech recognition. 463-466 - Gary Bradshaw, Alan Bell:
Towards the performance limits of connectionist feature detectors. 467-470 - Helge B. D. Sørensen:
Context-dependent and -independent self-structuring hidden control models for speech recognition. 471-474 - Marie-José Caraty, Claude Montacié, Claude Barras:
Integration of frequential and temporal structurations in a symbolic learning system. 475-478 - Enric Monte, José B. Mariño, Eduardo Lleida:
Smoothing hidden Markov models ay means of a self organizing feature map. 535-538 - Jyri Mäntysalo, Kari Torkkola, Teuvo Kohonen:
LVQ-based speech recognition with high-dimensional context vectors. 539-542 - Mikko Kurimo, Kari Torkkola:
Application of self-organizing maps and LVQ in training continuous density hidden Markov models for phonemes. 543-546 - Paul Dalsgaard, Ove Andersen:
Identification of mono- and poly-phonemes using acoustic-phonetic features derived by a self-organising neural network. 547-550 - Pekka Utela, Samuel Kaski, Kari Torkkola:
Using phoneme group specific LVQ-codebooks with HMMs. 551-554
Speech Synthesis 1-3
- Naoto Iwahashi, Yoshinori Sagisaka:
Speech segment network approach for an optimal synthesis unit set. 479-482 - Yoshinori Sagisaka, Nobuyoshi Kaiki, Naoto Iwahashi, Katsuhiko Mimura:
ATR μ-talk speech synthesis system. 483-486 - Bert Van Coile, Steven Leys, Luc Mortier:
On the development of a name pronunciation system. 487-490 - Inger Karlsson:
Consonants for female speech synthesis. 491-494 - Jan P. H. van Santen:
Diagnostic perceptual experiments for text-to-speech system evaluation. 555-558 - Marcello Balestri, Enzo Foti, Luciano Nebbia, Mario Oreglia, Pier Luigi Salza, Stefano Sandri:
Comparison of natural and synthetic speech intelligibility for a reverse telephone directory service. 559-562 - Richard Sproat, Julia Hirschberg, David Yarowsky:
A corpus-based synthesizer. 563-566 - Tomohisa Hirokawa, Kenzo Itoh, Hirokazu Sato:
High quality speech synthesis based on wavelet compilation of phoneme segments. 567-570 - David R. Williams, Corine A. Bickley, Kenneth N. Stevens:
Inventory of phonetic contrasts generated by high-level control of a formant synthesizer. 571-574 - Mikael Goldstein, Ove Till:
Is % overall error rate a valid measure of speech synthesiser and natural speech performance at the segmental level? 1131-1134 - Willy Jongenburger, Renée van Bezooijen:
Text-to-speech conversion for dutch: comprehensibility and acceptability. 1135-1138 - Masayo Katoh, Shin'ichiro Hashimoto:
The rhythm rules in Japanese based on the centers of energy gravity of vowels. 1139-1142 - Kenzo Itoh, Tomohisa Hirokawa, Hirokazu Sato:
Segmental power control for Japanese speech synthesis. 1143-1146 - Jean Schoentgen:
Glottal waveform synthesis with volterra shapers. 1147-1150 - Ken Ceder, Bertil Lyberg:
Yet another rule compiler for text-to-speech conversion? 1151-1154 - Kazuhiko Iwata, Yukio Mitome:
Prosody generation models constructed by considering speech tempo influence on prosody. 1155-1158 - Alex I. C. Monaghan:
Extracting microprosodic information from diphones - a simple way to model segmental effects on prosody for synthetic speech. 1159-1162 - Arjan van Hessen:
Generation of natural sounding speech stimuli by means of linear cepstral interpolation. 1163-1166 - W. Nick Campbell, Colin W. Wightman:
Prosodic encoding of syntactic structure for speech synthesis. 1167-1170 - Susan R. Hertz, Marie K. Huffman:
A nucleus-based timing model applied to multi-dialect speech synthesis by rule. 1171-1174 - Jill House, Nick J. Youd:
Evaluating the prosody of synthesized utterances within a dialogue system. 1175-1178 - Marcel Tatham, Eric Lewis:
Prosodics in a syllable-based text-to-speech synthesis system. 1179-1182 - Rabia Belrhali, Véronique Aubergé, Louis-Jean Boë:
From lexicon to rules: toward a descriptive method of French text-to-phonetics transcription. 1183-1186 - Marianne Elmlund, Ida Frehr, Niels Reinholt Petersen:
Formant transformation from male to female synthetic voices. 1187-1190 - Panagiotis A. Rentzepopoulos, George K. Kokkinakis:
Multilingual phoneme to grapheme conversion system based on HMM. 1191-1194 - Noriyo Hara, Hisayoshi Tsubaki, Hisashi Wakita:
Fundamental frequency control using linguistic information. 1195-1198 - Andrew P. Breen:
A comparison of statistical and rule based methods of determining segmental durations. 1199-1202 - J. R. Andrews, K. Mervyn Curtis, Volker Kraft:
Generation and extraction of high quality synthesis units. 1203-1206 - T. I. Boogaart, Kim E. A. Silverman:
Evaluating the overall comprehensibility of speech synthesizers. 1207-1210 - Olivier Boëffard, Laurent Miclet, S. White:
Automatic generation of optimized unit dictionaries for text to speech synthesis. 1211-1214 - Hideki Kasuya, Seiki Kasuya:
Relationships between syllable, word and sentence intelligibilities of synthetic speech. 1215-1218 - David R. Hill, Craig-Richard Schock, Leonard C. Manzara:
Unrestricted text-to-speech revisited: rhythm and intonation. 1219-1222 - Anton J. Rozsypal:
Wavelet speech synthesizer in the classroom and speech laboratory. 1223-1226 - Thomas Portele, Birgit Steffan, Rainer Preuß, Walter F. Sendlmeier, Wolfgang Hess:
HADIFIX - a speech synthesis system for German. 1227-1230 - Cristina Delogu, Stella Conte, Andrea Paoloni, Ciro Sementina:
Two different methodologies for evaluating the comprehension of synthetic passages. 1231-1234 - Carlos Gussenhoven, Toni C. M. Rietveld:
A target-interpolation model for the intonation of dutch. 1235-1238
Speech Perception: Phonetic Processes 1, 2
- Katharine Davis, Patricia K. Kuhl:
Best exemplars of English velar stops: a first report. 495-498 - Kenneth N. Stevens, Sharon Y. Manuel, Stefanie Shattuck-Hufnagel, Sharlene Liu:
Implementation of a model for lexical access based on features. 499-502 - Dieter Huber:
Perception of aperiodic speech signals. 503-506 - Hiroaki Kato, Minoru Tsuzaki, Yoshinori Sagisaka:
Acceptability and discrimination threshold for distortion of segmental duration in Japanese words. 507-510 - Anne Bonneau, Sylvie Coste, Linda Djezzar, Yves Laprie:
Two level acoustic cues for consistent stop identification. 511-514 - Rolf Carlson, James R. Glass:
Vowel classification based on analysis-by-synthesis. 575-578 - Maria-Gabriella Di Benedetto, Jean-Sylvain Liénard:
Extrinsic normalization of vowel formant values based on cardinal vowels mapping. 579-582 - Terrance M. Nearey:
Applications of generalized linear modeling to vowel data. 583-586 - David B. Pisoni:
Some comments on invariance, variability and perceptual normalization in speech perception. 587-590 - Stephen D. Goldinger, Thomas J. Palmeri, David B. Pisoni:
Words and voices: perceptual details are preserved in lexical representations. 591-594
Speech Enhancement
- Yan Ming Cheng, Douglas D. O'Shaughnessy, Peter Kabal:
Speech enhancement using a statistically derived filter mapping. 515-518 - V. Beattie, Steve J. Young:
Hidden Markov model state-based cepstral noise compensation. 519-522 - Guy J. Brown, Martin P. Cooke:
A computational model of auditory scene analysis. 523-526 - Srinivas Nandkumar, John H. L. Hansen, Robert J. Stets:
A new dual-channel speech enhancement technique with application to CELP coding in noise. 527-530 - Asunción Moreno, José A. R. Fonollosa:
CUMULANT - based voicing decision in noise corrupted speech. 531-534
Speech Recognition 1, 2
- Yolande Anglade, Dominique Fohr, Jean-Claude Junqua:
Selectively trained neural networks for the discrimination of normal and lombard speech. 595-598 - Aaron E. Rosenberg, Joel DeLong, Chin-Hui Lee, Biing-Hwang Juang, Frank K. Soong:
The use of cohort normalized scores for speaker verification. 599-602 - Tomoko Matsui, Sadaoki Furui:
Speaker recognition using concatenated phoneme models. 603-606 - Younès Bennani:
Speaker identification through a modular connectionist architecture: evaluation on the timit database. 607-610 - Claude Montacié, Jean-Luc Le Floch:
AR-vector models for free-text speaker recognition. 611-614 - Florian Schiel:
Rapid non-supervised speaker adaptation of semicontinuous hidden Markov models. 1463-1466 - D. Ederveen, Louis Boves:
Rule-based recognition of phoneme classes. 1467-1470 - Jie Yi, Kei Miki:
A new method of speaker-independent speech recognition using multiphone HMM. 1471-1474 - Myoung-Wan Koo, Chong Kwan Un:
A speaker adaptation based on corrective training and learning vector quantization. 1475-1478 - Katsuhiko Shirai, Shigeki Okawa, Tetsunori Kobayashi:
Phoneme recognition in continuous speech based on mutual information considering phonemic duration and connectivity. 1479-1482 - Shinji Koga, Ryosuke Isotani, Satoshi Tsukada, Kazunaga Yoshida, Kaichiro Hatazaki, Takao Watanabe:
A real-time speaker-independent continuous speech recognition system based on demi-syllable units. 1483-1486 - Saeed Vaseghi, Ben P. Milner:
Speech recognition in noisy environments. 1487-1490 - Fergus R. McInnes:
An enhanced interpolation technique for context-specific probability estimation in speech and language modelling. 1491-1494 - Lorenzo Fissore, Pietro Laface, Giorgio Micca, G. Sperto:
Channel adaptation for a continuous speech recognizer. 1495-1498 - S. Cifuentes, José Colás, Mohammad Hasan Savoji, José Manuel Pardo:
A new algorithm for connected digit recognition. 1499-1502 - Günther Ruske, Bernd Plannerer, Tanja Schultz:
Stochastic modeling of syllable-based units for continuous speech recognition. 1503-1506 - David M. Goblirsch, Toffee A. Albina:
HARK: an experimental speech recognition system. 1507-1510 - Akito Nagai, Jun-ichi Takami, Shigeki Sagayama:
The SSS-LR continuous speech recognition system: integrating SSS-derived allophone models and a phoneme-context-dependent LR parser. 1511-1514 - Shinsuke Sakai, Michael S. Phillips:
J-SUMMIT: a Japanese segment-based speech recognition system. 1515-1518 - Shinobu Mizuta, Kunio Nakajima:
Optimal discriminative training for HMMs to recognize noisy speech. 1519-1522 - Shingo Kuroiwa, Kazuya Takeda, Fumihiro Yato, Seiichi Yamamoto, Kunihiko Owa, Makoto Shozakai, Ryuuji Matsumoto:
Architecture and algorithms of a real-time word recognizer for telephone input. 1523-1526 - Hiroyasu Kuwano, Kazuya Nomura, Atsushi Ookumo, Shoji Hiraoka, Taisuke Watanabe, Katsuyuki Niyada:
Speaker independent speech recognition method using word spotting technique and its application to VCR programming. 1527-1530 - S. Lennon, Eliathamby Ambikairajah:
Transputer implementation of front-end processors for speech recognition systems. 1531-1534 - Yasuhiro Minami, Tatsuo Matsuoka, Kiyohiro Shikano:
Phoneme HMM evaluation algorithm without phoneme labeling. 1535-1538 - Andreas Noll, Henning Bergmann, Hans-Hermann Hamer, Annedore Paeseler, Horst Tomaschewski:
Architecture of a configurable application interface for speech recognition systems. 1539-1542 - Mark A. Fanty, John Pochmara, Ronald A. Cole:
An interactive environment for speech recognition research. 1543-1546 - Yoshiharu Abe, Kunio Nakajima:
An approach to unlimited vocabulary continuous speech recognition based on context-dependent phoneme modeling. 1547-1550 - Chuck Wooters, Nelson Morgan:
Acoustic subword models in the berkeley restaurant project. 1551-1554 - Claus Nedergaard Jacobsen:
SIRtrain, an open standard environment for CHMM recognizer development. 1555-1558 - Yutaka Kobayashi, Yasuhisa Niimi:
Segmented trellis algorithms for the continuous speech recognition. 1559-1562 - Bo Xu, Zhiwei Lin, Taiyi Huang, Dongxin Xu, Yuqing Gao:
A. 46 500 word Chinese speech recognition system. 1563-1566 - Dao Wen Chen:
Study of the time extension flat net for speech recognition. 1567-1568
Hidden Markov Models
- Frank Fallside:
A hidden Markov model structure for the acquisition of speech by machine, ASM. 615-618 - Yasuyuki Masai, Shin'ichi Tanaka, Tsuneo Nitta:
Speaker-independent keyword recognition based on SMQ/HMM. 619-622 - Régis Cardin, Diane Goupil, Roxane Lacouture, Evelyne Millien, Charles Snow, Yves Normandin:
CRIM's spontaneous speech recognition system for the ATIS task. 623-626 - Fabio Brugnara, Renato de Mori, Diego Giuliani, Maurizio Omologo:
Improved connected digit recognition using spectral variation functions. 627-630 - Andrew Tridgell, J. Bruce Millar, Kim-Anh Do:
Alternative preprocessing techniques for discrete hidden Markov model phoneme recognition. 631-634
Dialogue
- Gerhard Th. Niedermair:
Linguistic modelling in the context of oral dialogue. 635-638 - François Andry:
Static and dynamic predictions : a method to improve speech understanding in cooperative dialogues. 639-642 - Paul Heisterkamp, Scott McGlashan, Nick J. Youd:
Dialogue Semantics for an Oral Dialogue System. 643-646 - Masaaki Nagata:
Using pragmatics to rule out recognition errors in cooperative task-oriented dialogues. 647-650 - Yoichi Takebayashi, Hiroyuki Tsuboi, Yoichi Sadamoto, Hideki Hashimoto, Hideaki Shinchi:
A real-time speech dialogue system using spontaneous speech understanding. 651-654
Prosody and Phonology 1, 2
- Li-chiung Yang:
A semantic and pragmatic analysis of tone and intonation in Mandarin Chinese. 655-658 - Yoshimasa Tsukuma:
On prosodic features in speech - comparative studies between Japanese and standard Chinese. 659-662 - W. Nick Campbell:
Prosodic encoding of English speech. 663-666 - Gunnar Fant, Anita Kruckenberg, Lennart Nord:
Prediction of syllable duration, speech rate and tempo. 667-670 - Rolf Carlson, Björn Granström, Lennart Nord:
Experiments with emotive speech - acted utterances and synthesized replicas. 671-674 - J. Caspars, Vincent J. van Heuven:
Phonetic properties of dutch accent lending pitch movements under time pressure. 731-734 - Jacques M. B. Terken, Karin van den Hombergh:
Judgments of relative prominence for adjacent and non-adjacent accents. 735-738 - Frédéric Beaugendre, Christophe d'Alessandro, Anne Lacheret-Dujour, Jacques M. B. Terken:
A perceptual study of French intonation. 739-742 - Mark Liberman, J. Michael Schultz, Soonhyun Hong, Vincent Okeke:
The phonetics of IGBO tone. 743-746 - Stefanie Shattuck-Hufnagel:
Stress shift as pitch accent placement: within-word early accent placement in american English. 747-750
Speaking Styles 1, 2
- Katherine Morton:
Adding emotion to synthetic speech dialogue systems. 675-678 - Cari Spring, Donna Erickson, Thomas Call:
Emotional modalities and intonation in spoken language. 679-682 - Tatiana Slama-Cazacu:
Are any "press-conferences", "interviews" or "dialogues" true dialogues? 683-686 - Arthur J. Bronstein:
The categorization of the dialects and speech styles of north american English. 687-690 - Eleonora Blaauw:
Phonetic differences between read and spontaneous speech. 751-754 - Maxine Eskénazi:
Changing speech styles: strategies in read speech and casual and careful spontaneous speech. 755-758 - Noriko Umeda, Karen Wallace, Josephine Horna:
Usage of words and sentence structures in spontaneous versus text material. 759-762 - Nancy A. Daly, Victor Zue:
Statistical and linguistic analyses of F0 in read and spontaneous speech. 763-766 - Linda Shockey, Edda Farnetani:
Spontaneous speech in English and Italian. 767-770
ASR in Noise
- Claude Lefèbvre, Dariusz A. Zwierzynski, David R. Starks, Gary Birch:
Further optimisation of a robust IMELDA speech recogniser for applications with severely degraded speech. 691-694 - Richard M. Stern, Fu-Hua Liu, Yoshiaki Ohshima, Thomas M. Sullivan, Alejandro Acero:
Multiple approaches to robust speech recognition. 695-698 - Tadashi Kitamura, Satoshi Ando, Etsuro Hayahara:
Speaker-independent spoken digit recognition in noisy environments using dynamic spectral features and neural networks. 699-702 - Douglas A. Cairns, John H. L. Hansen:
ICARUS: an mwave-based real-time speech recognition system in noise and lombard effect. 703-706 - Chafic Mokbel, Laurent Barbier, Y. Kerlou, Gérard Chollet:
Word recognition in the car: adapting recognizers to new environments. 707-710
Dialogues and Applications
- Peter Meyer, Hans-Wilhelm Rühl, L. Vogten:
German announcements using synthetic speech the Gauss system. 711-714 - Mervyn A. Jack, John C. Foster, Fred Stentiford:
Intelligent dialogues in automated telephone services. 715-718 - Palle Bach Nielsen, Anders Baekgaard:
Experience with a dialogue description formalism for realistic applications. 719-722 - Solomon Lerner, Baruch Mazor:
Compensating for additive-noise in automatic speech recognition. 723-726 - Shoichi Matsunaga, Toshiaki Tsuboi, Tomokazu Yamada, Kiyohiro Shikano:
Continuous speech recognition for medical diagnoses using a character trigram model. 727-730
Sampling of Speech Production 1, 2
- Chengxiang Lu, Takayoshi Nakai, Hisayoshi Suzuki:
A three-dimensional FEM simulation of the effects of the vocal tract shape on the transfer function. 771-774 - Kiyoshi Oshimat, Vincent L. Gracco:
Mandibular contributions to speech production. 775-778 - Masafumi Matsumura:
Measurement of three-dimensional shapes of vocal tract and nasal cavity using magnetic resonance imaging technique. 779-782 - Shigeru Kiritani, Hajime Hirose, Kikuo Maekawa, Tsutomu Sato:
Electromyographie studies on the production of pitch contour in accentless dialects in Japanese. 783-786 - Yorinobu Sonoda, Kohichi Ogata:
Improvements of magnetometer sensing system for monitoring tongue point movements during speech. 843-846 - Paavo Alku:
Inverse filtering of the glottal waveform using the Itakura-saito distortion measure. 847-850 - Kunitoshi Motoki, Nobuhiro Miki:
Measurement of intraoral sound pressure distributions of Japanese vowels. 851-854
Labelling of Speech 1, 2
- Alain Marchal, William J. Hardcastle, K. Nicolaidis, Noël Nguyen, Fiona Gibbon:
Non-linear annotation of multi-channel speech data. 787-790 - Shingo Fujiwara, Yasuhiro Komori, Masahide Sugiyama:
A phoneme labelling workbench using HMM and spectrogram reading knowledge. 791-794 - Michael S. Phillips, Victor Zue:
Automatic discovery of acoustic measurements for phonetic classification. 795-798 - Katunobu Itou, Satoru Hayamizu, Hozumi Tanaka:
Detection of unknown words and automatic estimation of their transcriptions in continuous speech recognition. 799-802 - Fabio Brugnara, Daniele Falavigna, Maurizio Omologo:
A HMM-based system for automatic segmentation and labeling of speech. 803-806 - Robert W. P. Luk, Robert I. Damper:
A modification of the viterbi algorithm for stochastic phonographic transduction. 855-858 - Paul C. Bagshaw, Briony J. Williams:
Criteria for labelling prosodic aspects of English speech. 859-862 - Yifan Gong, Jean Paul Haton:
DTW-based phonetic labeling using explicit phoneme duration constraints. 863-866 - Kim E. A. Silverman, Mary E. Beckman, John F. Pitrelli, Mari Ostendorf, Colin W. Wightman, Patti Price, Janet B. Pierrehumbert, Julia Hirschberg:
TOBI: a standard for labeling English prosody. 867-870 - Barbara Eisen, Hans-Günther Tillmann, Christoph Draxler:
Consistency of judgements in manual labelling of phonetic segments: the distinction between clear and unclear cases. 871-874
Speech Production: Models and Theories 1, 2
- Gunnar Fant:
Vocal tract area functions of Swedish vowels and a new three-parameter model. 807-810 - Jean-Claude Junqua:
Acoustic and production pilot studies of speech vowels produced in noise. 811-814 - Yves Laprie, Marie-Odile Berger:
Active models for regularizing formant trajectories. 815-818 - René Carré, Samir Chennoukh, Mohamad Mrayati:
Vowel-consonant-vowel transitions: analysis, modeling, and synthesis. 819-822 - Maureen Stone, Subhash Lele:
Representing the tongue surface with curve fits. 875-878 - Katherine S. Harris, Eric Vatikiotis-Bateson, Peter J. Alfonso:
Muscle forces in vowel vocal tract formation. 879-882 - Makoto Hirayama, Eric Vatikiotis-Bateson, Mitsuo Kawato, Kiyoshi Honda:
Neural network modeling of speech motor control. 883-886 - Eric Vatikiotis-Bateson, Makoto Hirayama, Kiyoshi Honda, Mitsuo Kawato:
The articulatory dynamics of running speech: gestures from phonemes? 887-890
Phonetics
- Patricia A. Keating, B. Blankenship, Dani Byrd, Edward Flemming, Yuichi Todaka:
Phonetic analyses of the TIMIT corpus of american English. 823-826 - Dani Byrd:
Sex, dialects, and reduction. 827-830 - Manjari Ohala, John J. Ohala:
Phonetic universals and hindi segment duration. 831-834 - Donna Erickson, Osamu Fujimura:
Acoustic and articulatory correlates of contrastive emphasis in repeated corrections. 835-838 - Gary N. Tajchman, Marcia A. Bush:
Effects of context and redundancy in the perception of naturally produced English vowels. 839-842
Speech Data Bases
- Ronald A. Cole, Krist Roginski, Mark A. Fanty:
A telephone speech database of spelled and spoken names. 891-894 - Yeshwant K. Muthusamy, Ronald A. Cole, Beatrice T. Oshika:
The OGI multi-language telephone speech corpus. 895-898 - Douglas B. Paul, Janet M. Baker:
The design for the wall street journal-based CSR corpus. 899-902 - Lynette Hirschman:
Multi-site data collection for a spoken language corpus - MAD COW. 903-906 - Michael S. Phillips, James R. Glass, Joseph Polifroni, Victor Zue:
Collection and analyses of WSJ-CSR corpus at MIT. 907-910
Stochastic ASR
- Victor Abrash, Horacio Franco, Michael Cohen, Nelson Morgan, Yochai Konig:
Connectionist gender adaptation in a hybrid neural network / hidden Markov model speech recognition system. 911-914 - Michael Cohen, Horacio Franco, Nelson Morgan, David E. Rumelhart, Victor Abrash:
Hybrid neural network/hidden Markov model continuous-speech recognition. 915-918 - Gernot A. Fink, Franz Kummert, Gerhard Sagerer, Ernst Günter Schukat-Talamazzini, Heinrich Niemann:
Semantic hidden Markov networks. 919-922
Analysis of Disfluencies in Speech 1, 2
- Use Lehiste, Donna Erickson:
Hesitation sounds: is there coarticulation across pause? 923-926 - Corine A. Bickley, Sheri Hunnicutt:
Acoustic analysis of laughter. 927-930 - Douglas D. O'Shaughnessy:
Analysis of false starts in spontaneous speech. 931-934 - Robin J. Lickley, Ellen Gurman Bard:
Processing disfluent speech: recognising disfluency before lexical access. 935-938 - Philippe Morin, Jean-Claude Junqua, Jean-Marie Pierrel:
A flexible multimodal dialogue architecture independent of the application. 939-942 - Catia Cucchiarini, Renée van Bezooijen:
Familiarity with the language transcribed and context as determinants of intratranscriber agreement. 987-990 - Elizabeth Shriberg, Robin J. Lickley:
Intonation of clause-internal filled pauses. 991-994 - Elizabeth Wade, Elizabeth Shriberg, Patti Price:
User behaviors affecting speech recognition. 995-998
Linguistic Phonetics: Reduction
- Sharon Y. Manuel, Stefanie Shattuck-Hufnagel, Marie K. Huffman, Kenneth N. Stevens, Rolf Carlson, Sheri Hunnicutt:
Studies of vowel and consonant reduction. 943-946 - Noriko Umeda:
Formant frequencies of vowels in English function words. 947-950
Hearing/Speech Impaired 1, 2
- Christian Benoît, Tayeb Mohamadi:
The lip benefit: auditory and visual intelligibility of French speech in noise. 951-954 - Anne-Marie Öster:
Phonological assessment of deaf children's productive knowledge as a basis for speech-training. 955-958 - Hideaki Seki, Akiko Hayashi, Satoshi Imaizumi, Takehiko Harada, Hiroshi Hosoi:
Factors affecting voicing distinction of stops for the hearing impaired. 959-962 - Arthur Boothroyd, Robin S. Waldstein, Eddy Yeung:
Investigations into the auditory F0 speechreading enhancement effect using a sinusoidal replica of the F0 contour. 963-966 - Francesco Cutugno:
Some considerations on pitch and timing control in deaf children. 967-970 - Shari R. Baum:
Rate of speech effects in aphasia: an acoustic analysis of voice onset time. 1019-1022 - Parth M. Bhatt:
Fundamental frequency attributes following unilateral left or right temporal lobe lesion. 1023-1026 - Hiroshi Hosoi, Satoshi Imaizumi, Akiko Hayashi, Takehiko Harada, Hideaki Seki:
Cue extraction and integration in speech perception for the hearing impaired. 1027-1030 - Anna K. Nabelek:
The relationship between spectral details in naturally produced vowels and identification errors in noise and reverberation. 1031-1034 - Donald G. Jamieson, Leonard Cornelisse:
Speech processing effects on intelligibility for hearing-impaired listeners. 1035-1038
Unit-Based ASR
- Fuyuan Mo, Changli Li, Tao Chen:
Chinese recognition and synthesis system based on Chinese syllables. 971-974 - Hirofumi Yogo, Naoki Inagaki:
Accelerated stochastic approximation method based parameter estimation of monosyllables and their recognition using a neural network. 975-978 - Toomas Altosaar, Matti Karjalainen:
Diphone-based speech recognition using time-event neural networks. 979-982 - Giovanni Flammia, Paul Dalsgaard, Ove Andersen, Børge Lindberg:
Segment based variable frame rate speech analysis and recognition using a spectral variation function. 983-986
Language and Dialect Characterization
- Christian Benoît:
Intelligibility of the French spoken in France compared across listeners from France and from the Ivory Coast. 999-1002 - Julie Brousseau, Sally Anne Fox:
Dialect-dependent speech recognizers for canadian and european French. 1003-1006 - Yeshwant K. Muthusamy, Ronald A. Cole:
Automatic segmentation and identification of ten languages using telephone speech. 1007-1010 - Seiichi Nakagawa, Yoshio Ueda, Takashi Seino:
Speaker-independent, text-independent language identification by HMM. 1011-1014 - Shuichi Itahashi, Tsutomu Yamashita:
A discrimination method between Japanese dialects. 1015-1018
Speech Production, Parception, and Analysis
- B. Boyanov, Gérard Chollet:
Pathological voice analysis using cepstra, bispectra and group delay functions. 1039-1042 - Qianje Fu, Peyu Xia, Ren-Hua Wang:
Lateralization of speech sounds by binaural distributing processing. 1043-1046 - H. H. Rump:
Timing of pitch movements and perceived vowel duration. 1047-1050 - J. P. Liu, Geneviève Baudoin, Gérard Chollet:
Studies of glottal excitation and vocal tract parameters using inverse filtering and a parameterized input model. 1051-1054 - Dennis Norris, Brit van Ooyen, Anne Cutler:
Speeded detection of vowels and steady-state consonants. 1055-1058 - Elzbieta B. Slawinski:
Temporal factors in the perception of consonants for different age and hearing impairment groups. 1059-1062 - Abeer Alwan:
The role of F3 and F4 in identifying place of articulation for stop consonants. 1063-1066 - Thomas R. Sawallis:
A new measure for perceptual weight of acoustic cues: an experiment on voicing in French intervocalic [t, d]. 1067-1070 - Alan Wrench, Mervyn A. Jack, John Laver, Mary S. Jackson, David S. Soutar, A. Gerry Robertson, Janet MacKenzie Beck:
Objective speech quality assessment in patients with intra-oral cancers: voiceless fricatives. 1071-1074 - Bruce Connell:
Tongue contact, active articulators, and coarticulation. 1075-1078 - Makio Kashino, Astrid van Wieringen, Louis C. W. Pols:
Cross-languages differences in the identification of intervocalic stop consonants by Japanese and dutch listeners. 1079-1082 - Minoru Tsuzaki:
Effects of typicality and interstimulus interval on the discrimination of speech stimuli: within-subject comparison. 1083-1086 - Ronald A. Cole, Yeshwant K. Muthusamy:
Perceptual studies on vowels excised from continuous speech. 1087-1090 - Raymond S. Weitzman:
The relative perceptual salience of spectral and durational differences. 1095-1098 - Florien J. Koopmans-van Beinum:
Can 'level words' from one speaking style become teaks' when spliced into another speaking style? 1099-1102 - Beverley Gable, Helen Nemeth, Martin Haran:
Speech errors and task demand. 1103-1106 - John H. Esling, B. Craig Dickson, Roy C. Snell:
Analysis of phonation type using laryngographic techniques. 1107-1110 - Sumi Shigeno:
Effect of prototypes of vowels on speech perception in Japanese and English. 1111-1114 - Tomo-o Morohashi, Tetsuya Shimamura, Hiroyuki Yashima, Jouji Suzuki:
Characteristics of voice picked up from outer skin of larynx. 1115-1118 - Igor V. Nabelek:
Coding of voicing in whispered plosives. 1119-1122 - Margaret F. Cheesman, Shelly Lawrence, Allison Appleyard:
Performance on a nonsense syllable test using the articulation index. 1123-1126 - Donald G. Jamieson, Ketan Ramji, Issam Kheirallah, Terrance M. Nearey:
CSRE: a speech research environment. 1127-1130
Linguistic Phonetics
- Kazue Hata, Yoko Hasegawa:
A study of F0 reset in naturally-read utterances in Japanese. 1239-1242 - H. Samuel Wang, Fu-Dong Chiu:
On the nature of tone sandhi rules in taiwanese. 1243-1246 - Geoffrey S. Nathan:
How shallow is phonology: declarative phonologies meet fast speech. 1247-1250 - Junko Hosaka, Toshiyuki Takezawa, Noriyoshi Uratani:
Analyzing postposition drops in spoken Japanese. 1251-1254 - Jialu Zhang, Xinghui Hu:
Fundamental frequency patterns of Chinese in different speech modes. 1255-1258 - Knut Kvale, Arne Kjell Foldvik:
The multifarious r-sound. 1259-1262 - Zita McRobbie-Utasi:
The role of preaspiration duration in the voicing contrast in skolt sami. 1263-1266 - Eiji Yamada:
Parameter setting for abstract stress in tokyo Japanese. 1267-1270 - Georg Ottesen:
A method for studying prosody in texts read aloud. 1271-1274 - Vincent J. van Heuven:
Linguistic versus phonetic explanation of consonant lengthening after short vowels: a contrastive study of dutch and English. 1275-1278 - Kjell Elenius, Mats Blomberg:
Comparing phoneme and feature based speech recognition using artificial neural networks. 1279-1282 - Eva Strangert:
Prosodic cues to the perception of syntactic boundaries. 1283-1286 - Paul Taylor, Stephen Isard:
A new model of intonation for use with speech synthesis and recognition. 1287-1290 - Rudolf Weiss:
Computerized error detection/correction in teaching German sounds: some problems and solutions. 1291-1294 - Ahmed M. Elgendy:
Velum and epiglottis behavior during the production of Arabic pharyngeals and laryngeals: a fiberscopic study. 1295-1298 - Kim E. A. Silverman, Eleonora Blaauw, Judith Spitz, John F. Pitrelli:
A prosodic comparison of spontaneous speech and read speech. 1299-1302 - John J. Ohala, Maria Grazia Busà, Karen Harrison:
Phonological and psychological evidence that listeners normalize the speech signal. 1303-1306 - Elizabeth A. Hinkelman:
Intonation and the request/question distinction. 1307-1310 - Robert F. Port, Fred Cummins:
The English voicing contrast as velocity perturbation. 1311-1314 - Michael S. Ziolkowski, Mayumi Usami, Karen L. Landahl, Brenda K. Tunnock:
How many phonologies are there in one speaker? some experimental evidence. 1315-1318 - Hirokazu Sato:
Decomposition into syllable complexes and the accenting of Japanese loanwords. 1319-1322 - Jianfen Cao:
Temporal structure in bisyllabic word frame: an evidence for relational invariance and variability from standard Chinese. 1323-1326 - Shih-ping Wang:
The integration of phonetics and phonology: a case study of taiwanese "gemination" and syllable structure. 1327-1330
Perception and Production
- James H. Bradford:
Towards a robust speech interface for teleoperation systems. 1331-1334 - Piero Cosi, Paolo Frasconi, Marco Gori, N. Griggio:
Phonetic recognition experiments with recurrent neural networks. 1335-1338 - Mikael Goldstein, Björn Lindström, Ove Till:
Some aspects on context and response range effects when assessing naturalness of Swedish sentences generated by 4 synthesiser systems. 1339-1342 - Marcello Pelillo, Franca Moro, Mario Refice:
Probabilistic prediction of parts-of-speech from word spelling using decision trees. 1343-1346 - Dieter Barschdorff, Ulrich Gärtner:
Single word detection system with a neural classifier for recognizing speech at variable levels of background noise. 1347-1350 - Sharon L. Oviatt, Philip R. Cohen, Martin Fong, Michael Frank:
A rapid semi-automatic simulation technique for investigating interactive speech and handwriting. 1351-1354 - Sang-Hwa Chung, Dan I. Moldovan:
Speech understanding on a massively parallel computer. 1355-1358 - Chan-Do Lee:
Rationale for "performance phonology". 1359-1362 - Takuya Koizumi, Jyoji Urata, Shuji Taniguchi:
The effect of information feedback on the performance of a phoneme recognizer using kohonen map. 1363-1366 - Yasuharu Asano, Keikichi Hirose, Hiroya Fujisaki:
A method of dialogue management for the speech response system. 1367-1370 - Yumi Takizawa, Eiichi Tsuboka:
Syllable duration prediction for speech recognition. 1371-1374 - Jun He, Henri Leich:
Extracting fuzzy features from MLP for recognition of speech. 1379-1382 - Keiji Fukuzawa, Yoshinaga Kato, Masahide Sugiyama:
A fuzzy partition model (FPM) neural network architecture for speaker-independent continuous speech recognition. 1383-1386 - A. Ennaji, Jean Rouat:
Conception of speech filters based on a neural network. 1387-1390 - Hong-Kwang Jeff Kuo, Chin-Hui Lee, Aaron E. Rosenberg:
Speaker set identification through speaker group modeling. 1391-1394 - Stephen Springer, Sara Basson, Judith Spitz:
Identification of principal ergonomic requirements for interactive spoken language systems. 1395-1398 - Thomas E. Jacobs, Eric R. Buhrke:
Performance of the united kingdom intelligent network automatic speech recognition system. 1399-1402 - Guy Deville, Pierre Mousel:
Evaluation of parsing strategies in natural language spoken man-machine dialogue. 1403-1406 - Yasuhisa Niimi, Yutaka Kobayashi:
An information retrieval system with a speech interface. 1407-1410 - Julian P. Eatock, John S. D. Mason:
Phoneme performance in speaker recognition. 1411-1414 - Evelyne Tzoukermann, Roberto Pieraccini, Zakhar Gorelov:
Natural language processing in the chronus system. 1415-1418 - Dominique François, Dominique Fohr:
Contribution of neural networks for phoneme identification in the APHODEX expert system. 1419-1422 - Douglas B. Paul:
A CSR-NL interface architecture. 1423-1426 - Pierre Lefebvre, Frank Poirier, G. Duncan:
Speech interface for a man-machine dialog with the unix operating system. 1427-1430 - P. Bardaud, François Capman, Chafic Mokbel, Chakib Tadj, Gérard Chollet:
Transformation of databases for the evaluation of speech recognizers. 1431-1434 - Yoichi Yamashita, Riichiro Mizoguchi:
Dialog management for speech output from concept representation. 1435-1438 - Seiichiro Hangai, Shigetoshi Sugiyama, Kazuhiro Miyauchi:
Speaker verification using locations and sizes of multipulses on neural networks. 1439-1442 - Carlos Teixeira, Isabel Trancoso:
Word rejection using multiple sink models. 1443-1446 - Boerge Lindberg:
Verification of language specific performance factors from recogniser testing on EUROM.1 CVC material. 1447-1450 - Alain Cozannet:
Modeling task driven oral dialogue. 1451-1454 - Wei-ying Li, Kechu Yi, Zheng Hu:
Introducing neural predictor to hidden Markov model for speech recognition. 1455-1458 - Feng Liu, Jianxin Jiang, Jun Cheng, Kechu Yi:
A neural network based on subnets - SNN. 1459-1462
Analysis/Systems
- Ute Ziegenhain:
Syntactic anaphora resolution in a speech understanding system. 1569-1572 - Marion Mast, Ralf Kompe, Franz Kummert, Heinrich Niemann, Elmar Nöth:
The dialog module of the speech recognition and dialog system EVAR. 1573-1576 - Yan Ming Cheng, Douglas D. O'Shaughnessy, Paul Mermelstein:
Statistical recovery of wideband speech from narrowband speech. 1577-1580 - Henk van den Heuvel, Toni C. M. Rietveld:
Speaker related variability in cepstral representations of dutch speech segments. 1581-1584 - Per Rosenbeck, Bo Baungaard:
Experiences from a real-world telephone application: teledialogue. 1585-1588 - K. Y. Lee, P. Ha, J. Rheem, Souguil Ann, Iickho Song:
Robust estimation of time-varying LP parameters on speech. 1589-1592 - Javier Hernando, Climent Nadeu, Eduardo Lleida:
On the AR modelling of the one-sided autocorrelation sequence for noisy speech recognition. 1593-1596 - Hiroshi Shimodaira, Mitsuru Nakai:
Robust pitch detection by narrow band spectrum analysis. 1597-1600 - S. Eady, B. Craig Dickson, Roy C. Snell, J. Woolsey, P. Ollek, A. Wynrib, Jocelyn Clayards:
A microcomputer-based system for real-time analysis and display of laryngograph signals. 1601-1604 - Nanette M. Veilleux, Mari Ostendorf, Colin W. Wightman:
Parse scoring with prosodic information. 1605-1608 - Ying Cheng, Paul Fortier, Yves Normandin:
Topic identification using a neural network with a keyword-spotting preprocessor. 1609-1612 - Shane Switzer, Timothy R. Anderson, Matthew Kabrisky, Steven K. Rogers, Bruce W. Suter:
Frequency domain speech coding. 1613-1616 - Raymond Descout, Robert Bergeron, Bernard Meriald:
MEDIATEX-TASF: a closed captioning real-time service in French. 1617-1620 - S. A. Wilde, K. Mervyn Curtis:
The wavelet transform for speech analysis. 1621-1624 - Pablo Aibar, Andrés Marzal, Enrique Vidal, Francisco Casacuberta:
Problems and algorithms in optimal linguistic decoding: a unified formulation. 1625-1628 - Jean Rouat, Sylvain Lemieux, Alain Migneault:
A spectro-temporal analysis of speech based on nonlinear operators. 1629-1632 - Miguel A. Berrojo, Javier Corrales, Jesus Macias, Santiago Aguilera:
A PC graphic tool for speech research based on a DSP board. 1633-1636 - Satoru Hayamizu, Katunobu Itou, Masafumi Tamoto, Kazuyo Tanaka:
A spoken language dialogue system for automatic collection of spontaneous speech. 1637-1640 - Shingo Nishioka, Yoichi Yamashita, Riichiro Mizoguchi:
A powerful disambiguating mechanism for speech understanding systems based on ATMs. 1641-1644 - Najib Naja, Jean-Marc Boucher, Samir Saoudi:
A mixed Gaussian-stochastic code book for CELP coder in LSP speech coding. 1645-1648 - Hiroyuki Kamata, Yoshihisa Ishida:
A method to estimate the transfer function of ARMA model of speech wave using prony method and homomorphic analysis. 1649-1652 - Børge Lindberg, Bjarne Andersen, Anders Baekgaard, Tom Brøndsted, Paul Dalsgaard, Jan Kristiansen:
An integrated dialogue design and continuous speech recognition system environment. 1653-1656 - Alain Marchal, Christine Meunier, P. Gavarry:
The PSH/DISPE helium speech cdrom. 1657-1660
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.