Acoustic Phonetics 1

Home
Full List of Titles
1: ICSLP'98 Proceedings
Keynote Speeches
Text-To-Speech Synthesis 1
Spoken Language Models and Dialog 1
Prosody and Emotion 1
Hidden Markov Model Techniques 1
Speaker and Language Recognition 1
Multimodal Spoken Language Processing 1
Isolated Word Recognition
Robust Speech Processing in Adverse Environments 1
Spoken Language Models and Dialog 2
Articulatory Modelling 1
Talking to Infants, Pets and Lovers
Robust Speech Processing in Adverse Environments 2
Spoken Language Models and Dialog 3
Speech Coding 1
Articulatory Modelling 2
Prosody and Emotion 2
Neural Networks, Fuzzy and Evolutionary Methods 1
Utterance Verification and Word Spotting 1 / Speaker Adaptation 1
Text-To-Speech Synthesis 2
Spoken Language Models and Dialog 4
Human Speech Perception 1
Robust Speech Processing in Adverse Environments 3
Speech and Hearing Disorders 1
Prosody and Emotion 3
Spoken Language Understanding Systems 1
Signal Processing and Speech Analysis 1
Spoken Language Generation and Translation 1
Spoken Language Models and Dialog 5
Segmentation, Labelling and Speech Corpora 1
Multimodal Spoken Language Processing 2
Prosody and Emotion 4
Neural Networks, Fuzzy and Evolutionary Methods 2
Large Vocabulary Continuous Speech Recognition 1
Speaker and Language Recognition 2
Signal Processing and Speech Analysis 2
Prosody and Emotion 5
Robust Speech Processing in Adverse Environments 4
Segmentation, Labelling and Speech Corpora 2
Speech Technology Applications and Human-Machine Interface 1
Large Vocabulary Continuous Speech Recognition 2
Text-To-Speech Synthesis 3
Language Acquisition 1
Acoustic Phonetics 1
Speaker Adaptation 2
Speech Coding 2
Hidden Markov Model Techniques 2
Multilingual Perception and Recognition 1
Large Vocabulary Continuous Speech Recognition 3
Articulatory Modelling 3
Language Acquisition 2
Speaker and Language Recognition 3
Text-To-Speech Synthesis 4
Spoken Language Understanding Systems 4
Human Speech Perception 2
Large Vocabulary Continuous Speech Recognition 4
Spoken Language Understanding Systems 2
Signal Processing and Speech Analysis 3
Human Speech Perception 3
Speaker Adaptation 3
Spoken Language Understanding Systems 3
Multimodal Spoken Language Processing 3
Acoustic Phonetics 2
Large Vocabulary Continuous Speech Recognition 5
Speech Coding 3
Language Acquisition 3 / Multilingual Perception and Recognition 2
Segmentation, Labelling and Speech Corpora 3
Text-To-Speech Synthesis 5
Spoken Language Generation and Translation 2
Human Speech Perception 4
Robust Speech Processing in Adverse Environments 5
Text-To-Speech Synthesis 6
Speech Technology Applications and Human-Machine Interface 2
Prosody and Emotion 6
Hidden Markov Model Techniques 3
Speech and Hearing Disorders 2 / Speech Processing for the Speech and Hearing Impaired 1
Human Speech Production
Segmentation, Labelling and Speech Corpora 4
Speaker and Language Recognition 4
Speech Technology Applications and Human-Machine Interface 3
Utterance Verification and Word Spotting 2
Large Vocabulary Continuous Speech Recognition 6
Neural Networks, Fuzzy and Evolutionary Methods 3
Speech Processing for the Speech-Impaired and Hearing-Impaired 2
Prosody and Emotion 7
2: SST Student Day
SST Student Day - Poster Session 1
SST Student Day - Poster Session 2

Author Index
A B C D E F G H I
J K L M N O P Q R
S T U V W X Y Z

Multimedia Files

Assimilation of Place in Japanese and Dutch

Authors:

Anne Cutler, Max Planck Institute for Psycholinguistics (The Netherlands)
Takashi Otake, Dokkyo University (Japan)

Page (NA) Paper number 93

Abstract:

Assimilation of place of articulation across a nasal and a following stop consonant is obligatory in Japanese, but not in Dutch. In four experiments the processing of assimilated forms by speakers of Japanese and Dutch was compared, using a task in which listeners blended pseudo-word pairs such as ranga-serupa. An assimilated blend of this pair would be rampa, an unassimilated blend rangpa. Japanese listeners produced significantly more assimilated than unassimilated forms, both with pseudo-Japanese and pseudo-Dutch materials, while Dutch listeners produced significantly more unassimilated than assimilated forms in each materials set. This suggests that Japanese listeners, whose native-language phonology involves obligatory assimilation constraints, represent the assimilated nasals in nasal-stop sequences as unmarked for place of articulation, while Dutch listeners, who are accustomed to hearing unassimilated forms, represent the same nasal segments as marked for place of articulation.

SL980093.PDF (From Author) SL980093.PDF (Rasterized)

TOP


Prosodic Constraint on V-to-V Coarticulation in Japanese

Authors:

Yuko Kondo, Musashino Art University (Japan)
Yumiko Arai, Y & Y Phonetic Laboratory (Japan)

Page (NA) Paper number 186

Abstract:

Coarticulation is considered to be constrained by language specific grammar. In particular, prosodic units, which have an important function in organizing speech, are likely to affect the extent of coarticulation. The present study addresses the question of the relationship between V-to-V coarticulation and bimoraic foot in Japanese. It was shown that the foot boundary constrains the extent of coarticulation. It was also shown that both within- and across-foot, anticipatory effects are stronger than carryover effects in Japanese.

SL980186.PDF (From Author) SL980186.PDF (Rasterized)

TOP


Postvocalic /r/-deletion in Standard Dutch: How Experimental Phonology Can Profit From ASR Technology

Authors:

Catia Cucchiarini, A2RT, University of Nijmegen (The Netherlands)
Henk van den Heuvel, A2RT, University of Nijmegen (The Netherlands)

Page (NA) Paper number 753

Abstract:

In this study automatic speech recognition (ASR) techniques were used to substantiate the findings of phonological research on postvocalic /r/-deletion in Standard Dutch. A database containing spontaneous speech utterances stemming from man-machine interactions in an automatic train-table inquiry system was used for this purpose. Pronunciation variants with and without /r/ were automatically generated on the basis of our specification of the phonological rule of /r/-deletion and were then included in the lexicon of a continuous speech recognizer (CSR) which was used in forced recognition mode. The results show that in a corpus containing 214,102 words, in which /r/-deletion could be applied 16,865 times, it was actually applied in 47.6% of the cases. This is a high percentage of occurrence for a phenomenon that has so sporadically been described. Furthermore, the results substantiate our rule specification: /r/-deletion occurs more often after schwa than after any other vowel.

SL980753.PDF (From Author) SL980753.PDF (Rasterized)

TOP


More Evidence For The Perceptual Basis Of Sound Change? Suprasegmental Effects In The Development Of Distinctive Nasalization

Authors:

John Hajek, University of Melbourne (Australia)
Ian Watson, University of Oxford (U.K.)

Page (NA) Paper number 254

Abstract:

Cross-linguistic studies of the development of distinctive nasalization show evidence of significant suprasegmental conditioning. Amongst conditioning factors uncovered so far are vowel length and stress. Across languages it is reported that in the related contexts /V:N/ and /VN/, identical except for vowel length, phonologization of nasalization and N-deletion always occur preferentially in the context of long vowels. There is also cross-linguistic evidence of stress-conditioning of distinctive nasalization: nasalization and N-deletion appear to occur preferentially in stressed syllables. In this study, we discuss the results of an experiment designed to measure the possible effects of vowel duration and prominence on the perception of vowel nasalization. Both are seen to have an effect, although in different ways. Results presented here also lend support to the hypothesis that some sound changes, such as those involved in distinctive nasalization, may have a primarily perceptual basis.

SL980254.PDF (From Author) SL980254.PDF (Rasterized)

TOP


Speech Production Of Vowel Sequences Using A Physiological Articulatory Model

Authors:

Jianwu Dang, ATR Human Information Processing Research Labs (Japan)
Kiyoshi Honda, ATR Human Information Processing Research Labs (Japan)

Page (NA) Paper number 639

Abstract:

This report describes the development of a physiologically-based articulatory model, which consists of the tongue, mandible, hyoid bone and vocal tract wall. These organs are represented as a midsagittal quasi-3D layer with a thickness of 2 cm for tongue tissue and 3 cm for tract wall. The geometry of these organs and muscles are extracted from volumetric MR images of a male speaker. Both the soft and rigid structures are represented by mass-points and viscoelastic springs for connective tissue, where the springs for bony organs are set to extremely large stiffness. This design is suitable to compute soft tissue deformations and rigid organ displacements simultaneously using a single algorithm, and thus reduces computational complexities of the simulation. A novel control method is developed to produce dynamic actions of the vocal tract, as well as to handle the collision of the tongue to surrounding walls. Area functions are obtained for vowel sequences based on model's vocal tract widths in the midsagittal and parasagittal planes. The proposed model demonstrated plausible dynamic behaviors for human speech articulation.

SL980639.PDF (From Author) SL980639.PDF (Rasterized)

TOP