Language Acquisition 1

Home
Full List of Titles
1: ICSLP'98 Proceedings
Keynote Speeches
Text-To-Speech Synthesis 1
Spoken Language Models and Dialog 1
Prosody and Emotion 1
Hidden Markov Model Techniques 1
Speaker and Language Recognition 1
Multimodal Spoken Language Processing 1
Isolated Word Recognition
Robust Speech Processing in Adverse Environments 1
Spoken Language Models and Dialog 2
Articulatory Modelling 1
Talking to Infants, Pets and Lovers
Robust Speech Processing in Adverse Environments 2
Spoken Language Models and Dialog 3
Speech Coding 1
Articulatory Modelling 2
Prosody and Emotion 2
Neural Networks, Fuzzy and Evolutionary Methods 1
Utterance Verification and Word Spotting 1 / Speaker Adaptation 1
Text-To-Speech Synthesis 2
Spoken Language Models and Dialog 4
Human Speech Perception 1
Robust Speech Processing in Adverse Environments 3
Speech and Hearing Disorders 1
Prosody and Emotion 3
Spoken Language Understanding Systems 1
Signal Processing and Speech Analysis 1
Spoken Language Generation and Translation 1
Spoken Language Models and Dialog 5
Segmentation, Labelling and Speech Corpora 1
Multimodal Spoken Language Processing 2
Prosody and Emotion 4
Neural Networks, Fuzzy and Evolutionary Methods 2
Large Vocabulary Continuous Speech Recognition 1
Speaker and Language Recognition 2
Signal Processing and Speech Analysis 2
Prosody and Emotion 5
Robust Speech Processing in Adverse Environments 4
Segmentation, Labelling and Speech Corpora 2
Speech Technology Applications and Human-Machine Interface 1
Large Vocabulary Continuous Speech Recognition 2
Text-To-Speech Synthesis 3
Language Acquisition 1
Acoustic Phonetics 1
Speaker Adaptation 2
Speech Coding 2
Hidden Markov Model Techniques 2
Multilingual Perception and Recognition 1
Large Vocabulary Continuous Speech Recognition 3
Articulatory Modelling 3
Language Acquisition 2
Speaker and Language Recognition 3
Text-To-Speech Synthesis 4
Spoken Language Understanding Systems 4
Human Speech Perception 2
Large Vocabulary Continuous Speech Recognition 4
Spoken Language Understanding Systems 2
Signal Processing and Speech Analysis 3
Human Speech Perception 3
Speaker Adaptation 3
Spoken Language Understanding Systems 3
Multimodal Spoken Language Processing 3
Acoustic Phonetics 2
Large Vocabulary Continuous Speech Recognition 5
Speech Coding 3
Language Acquisition 3 / Multilingual Perception and Recognition 2
Segmentation, Labelling and Speech Corpora 3
Text-To-Speech Synthesis 5
Spoken Language Generation and Translation 2
Human Speech Perception 4
Robust Speech Processing in Adverse Environments 5
Text-To-Speech Synthesis 6
Speech Technology Applications and Human-Machine Interface 2
Prosody and Emotion 6
Hidden Markov Model Techniques 3
Speech and Hearing Disorders 2 / Speech Processing for the Speech and Hearing Impaired 1
Human Speech Production
Segmentation, Labelling and Speech Corpora 4
Speaker and Language Recognition 4
Speech Technology Applications and Human-Machine Interface 3
Utterance Verification and Word Spotting 2
Large Vocabulary Continuous Speech Recognition 6
Neural Networks, Fuzzy and Evolutionary Methods 3
Speech Processing for the Speech-Impaired and Hearing-Impaired 2
Prosody and Emotion 7
2: SST Student Day
SST Student Day - Poster Session 1
SST Student Day - Poster Session 2

Author Index
A B C D E F G H I
J K L M N O P Q R
S T U V W X Y Z

Multimedia Files

Non-Native Productions Of Japanese Single Stops That Are Too Long For One Mora Unit

Authors:

Yasuyo Minagawa-Kawai, University of Tokyo (Japan)
Shigeru Kiritani, University of Tokyo (Japan)

Page (NA) Paper number 1025

Abstract:

The difficulty for non-native speakers in producing Japanese geminate stops with a long enough closure has been pointed out and extensively studied. However, the reverse problem exists for particular language speakers such as Chinese. For these speakers, production of Japanese intervocalic single stops sound like geminate stops. This study aimed to show acoustic evidence of this problem in the production of Japanese voiceless stops by learners of Japanese, and to compare speakers from different language groups, English, Chinese and French speakers. The results of the experiment indicated that since some of the stop productions by Americans and Chinese were too long for one mora unit, the Japanese heard a geminate (2 mora) instead of a single. Because French subjects had a small difference in the stop durations from the Japanese value, only a few of their productions of single stops were perceived as geminates. The language differences in timing pattern for Japanese stops were interpreted to reflect differences in timing implementation of native languages that is stress-timed and syllable-timed, at least concerning the difference between English and French speakers.

SL981025.PDF (From Author) SL981025.PDF (Rasterized)

TOP


The Process Of Generation And Development Of Second Language Japanese Accentuation

Authors:

Nobuko Yamada, Faculty of Humanities, Ibaraki University (Japan)

Page (NA) Paper number 1056

Abstract:

This study will investigate how non-native speakers of Japanese acquire Japanese accentuation from the viewpoint of the location of the accent nucleus. Hypothetical models for the process of generation and for developmental sequence of interlanguage Japanese accentuation, which is interim accentual system created by learners, will be proposed. The subjects appear to generate their interlanguage as the results of application of strategies or examples of accentuation. Those seem to be discovered from L2 input, or chosen and fetch from their memory. The subjects' competence of accentuation appear to be developed by L2 input, starting with L1 and universal property. They seem to discover and apply 5 types of strategies toward acquisition of target accentual rules of Japanese.

SL981056.PDF (From Author) SL981056.PDF (Scanned)

TOP


Perceptual Properties of Russians with Japanese Fricatives

Authors:

Seiya Funatsu, Hiroshima Women's University (Japan)
Shigeru Kiritani, The University of Tokyo (Japan)

Page (NA) Paper number 809

Abstract:

This study investigated the perceptual properties of second language learners in acquiring second language phonemes. The case where the relation between two phonemes of a second language and those of a native language changes according to following vowels was studied. The perceptual properties of Russians with regards to Japanese fricatives were examined. In the perception test, the confusion of [(omitted)o] with [so] was very large. This phenomenon could be caused by the difference between the transition onset time from [s'] to vowels and that from the other consonants to vowels. It is considered that, in the case of following vowel [a] and [o], Russians equated Japanese [s] and [(omitted)] with Russian [s] and [s'] respectively. However, in the case of [u], they did not equate them in such a manner. This is probably because the acoustic properties of Japanese [(omitted)] are very different from those of Russian [u].

SL980809.PDF (From Author) SL980809.PDF (Rasterized)

TOP


Assessment of Dutch Pronunciation by Means of Automatic Speech Recognition Technology

Authors:

Catia Cucchiarini, A2RT, University of Nijmegen (The Netherlands)
Febe De Wet, A2RT, University of Nijmegen (The Netherlands)
Helmer Strik, A2RT, University of Nijmegen (The Netherlands)
Louis Boves, A2RT, University of Nijmegen (The Netherlands)

Page (NA) Paper number 751

Abstract:

Experiments were carried out to determine whether log-likelihood ratios (LRs) can be employed to improve automatic assessment of Dutch pronunciation. Read speech of natives and non-natives was judged by three groups of expert raters and was then analyzed by means of a continuous speech recognizer. Three automatic measures were calculated, two LRs and rate of speech (ros), and then compared with the expert ratings. It appears that expert ratings of pronunciation quality can accurately be predicted on the basis of ros alone and that LRs do not contribute to better prediction. However, LRs can be useful to automatic pronunciation assessment because they can help detect fast speakers who produce totally wrong sentences.

SL980751.PDF (From Author) SL980751.PDF (Rasterized)

TOP


Phonetic-Level Mispronunciation Detection in Non-Native Swedish Speech

Authors:

Philippe Langlais, CTT - KTH (Sweden)
Anne-Marie Öster, TMH-KTH (Sweden)
Björn Granström, CTT- KTH (Sweden)

Page (NA) Paper number 311

Abstract:

This contribution presents part of the work initiated at the CTT for the development of speech technology to assist non-native speakers learn Swedish. This study focuses mainly on the automatic location of mispronunciations at a phonetic level. We first describe the database we created for this work and then report on the reliability of several phonetic scores to automatically locate segmental problems in student utterances.

SL980311.PDF (From Author) SL980311.PDF (Rasterized)

TOP


Computer-Based Second Language Production Training By Using Spectrographic Representation And HMM-Based Speech Recognition Scores

Authors:

Reiko Akahane-Yamada, ATR Human Information Processing Research Laboratories (Japan)
Erik McDermott, ATR Human Information Processing Research Laboratories (Japan)
Takahiro Adachi, ATR Human Information Processing Research Laboratories (Japan)
Hideki Kawahara, ATR Human Information Processing Research Laboratories (Japan)
John S. Pruitt, ATR Human Information Processing Research Laboratories (Japan)

Page (NA) Paper number 429

Abstract:

How can we provide feedback to second language (L2) learners about the goodness of their productions in an automatic way? In this paper, we introduce our attempts to provide effective feedback when we train native speakers of Japanese to produce English /r/ and /l/. First, we adopted spectrographic representation overlayed with formant frequencies as feedback. Second, we investigated the correlation between human judgments of L2 production quality and acoustic scores produced by an HMM-based speech recognition system. We also adopted the HMM-based scores as feedback in the production training. Evaluation of the pre- and post-training productions by human judges showed that production abilities of the trainees improved in both training groups, suggesting that both spectrographic representation and HMM-based scores were useful and meaningful as feedback. These results are discussed in the context of optimizing L2 speech training.

SL980429.PDF (From Author) SL980429.PDF (Rasterized)

TOP