Speech Technology Applications and Human-Machine Interface 3

Home
Full List of Titles
1: ICSLP'98 Proceedings
Keynote Speeches
Text-To-Speech Synthesis 1
Spoken Language Models and Dialog 1
Prosody and Emotion 1
Hidden Markov Model Techniques 1
Speaker and Language Recognition 1
Multimodal Spoken Language Processing 1
Isolated Word Recognition
Robust Speech Processing in Adverse Environments 1
Spoken Language Models and Dialog 2
Articulatory Modelling 1
Talking to Infants, Pets and Lovers
Robust Speech Processing in Adverse Environments 2
Spoken Language Models and Dialog 3
Speech Coding 1
Articulatory Modelling 2
Prosody and Emotion 2
Neural Networks, Fuzzy and Evolutionary Methods 1
Utterance Verification and Word Spotting 1 / Speaker Adaptation 1
Text-To-Speech Synthesis 2
Spoken Language Models and Dialog 4
Human Speech Perception 1
Robust Speech Processing in Adverse Environments 3
Speech and Hearing Disorders 1
Prosody and Emotion 3
Spoken Language Understanding Systems 1
Signal Processing and Speech Analysis 1
Spoken Language Generation and Translation 1
Spoken Language Models and Dialog 5
Segmentation, Labelling and Speech Corpora 1
Multimodal Spoken Language Processing 2
Prosody and Emotion 4
Neural Networks, Fuzzy and Evolutionary Methods 2
Large Vocabulary Continuous Speech Recognition 1
Speaker and Language Recognition 2
Signal Processing and Speech Analysis 2
Prosody and Emotion 5
Robust Speech Processing in Adverse Environments 4
Segmentation, Labelling and Speech Corpora 2
Speech Technology Applications and Human-Machine Interface 1
Large Vocabulary Continuous Speech Recognition 2
Text-To-Speech Synthesis 3
Language Acquisition 1
Acoustic Phonetics 1
Speaker Adaptation 2
Speech Coding 2
Hidden Markov Model Techniques 2
Multilingual Perception and Recognition 1
Large Vocabulary Continuous Speech Recognition 3
Articulatory Modelling 3
Language Acquisition 2
Speaker and Language Recognition 3
Text-To-Speech Synthesis 4
Spoken Language Understanding Systems 4
Human Speech Perception 2
Large Vocabulary Continuous Speech Recognition 4
Spoken Language Understanding Systems 2
Signal Processing and Speech Analysis 3
Human Speech Perception 3
Speaker Adaptation 3
Spoken Language Understanding Systems 3
Multimodal Spoken Language Processing 3
Acoustic Phonetics 2
Large Vocabulary Continuous Speech Recognition 5
Speech Coding 3
Language Acquisition 3 / Multilingual Perception and Recognition 2
Segmentation, Labelling and Speech Corpora 3
Text-To-Speech Synthesis 5
Spoken Language Generation and Translation 2
Human Speech Perception 4
Robust Speech Processing in Adverse Environments 5
Text-To-Speech Synthesis 6
Speech Technology Applications and Human-Machine Interface 2
Prosody and Emotion 6
Hidden Markov Model Techniques 3
Speech and Hearing Disorders 2 / Speech Processing for the Speech and Hearing Impaired 1
Human Speech Production
Segmentation, Labelling and Speech Corpora 4
Speaker and Language Recognition 4
Speech Technology Applications and Human-Machine Interface 3
Utterance Verification and Word Spotting 2
Large Vocabulary Continuous Speech Recognition 6
Neural Networks, Fuzzy and Evolutionary Methods 3
Speech Processing for the Speech-Impaired and Hearing-Impaired 2
Prosody and Emotion 7
2: SST Student Day
SST Student Day - Poster Session 1
SST Student Day - Poster Session 2

Author Index
A B C D E F G H I
J K L M N O P Q R
S T U V W X Y Z

Multimedia Files

Is Speech The Right Thing For Your Application?

Authors:

Niels Ole Bernsen, Odense University (Denmark)
Laila Dybkjær, Odense University (Denmark)

Page (NA) Paper number 62

Abstract:

Use of speech input to, and speech output from, computer systems is spreading at a growing pace. This means that, increasingly, developers of systems and interfaces are faced with the question of whether or not to use speech input and/or speech output for the applications they are about to build. This paper presents results from a pilot test of a theory-based approach to speech functionality. The test uses a corpus of claims about speech functionality derived from recent literature on speech and multimodality.

SL980062.PDF (From Author) SL980062.PDF (Rasterized)

TOP


A PC-Based Tool for Helping in Diagnosis of Pathologic Voice

Authors:

Juan Ignacio Godino Llorente, ETSI Telecomunicacion (UPM) (Spain)
Santiago Aguilera Navarro, ETSI Telecomunicacion (UPM) (Spain)
Sira Palazuelos Cagigas, ETSI Telecomunicacion (UPM) (Spain)
Alberto Nieto Altuzarra, Universidad de Alcala de Henares (Spain)
Pedro Gómez Vilda, ETSI Telecomunicacion (UPM) (Spain)

Page (NA) Paper number 558

Abstract:

We present a diagnosis tool for the voice clinic which runs on a personal computer. The application records different registers in real time. The signals to be captured and stored are the following: * Videoendoscopic images recorded with fibroscope or telelaringoscope. * Electroglottographic signal during fonation * Voice signal * Air flow signal All these different signals are recorded with specific transducers and standard digitalisation signal boards, using microphone input and input line simultaneously. Several systems to help in the diagnostic have been previously developed, but none of them captures the four mentioned signals simultaneously. All of them are highly interesting from the clinical point of view, and assist the expert in making the decision. The main advantage, related to other systems, is that looking at the videoendoscopic record, clinicians are able to label voice registers with the associated pathologies.

SL980558.PDF (From Author) SL980558.PDF (Rasterized)

TOP


Web-Based Educational Tools for Speech Technology

Authors:

Kaare Sjölander, KTH (Sweden)
Jonas Beskow, KTH (Sweden)
Joakim Gustafson, KTH (Sweden)
Erland Lewin, KTH (Sweden)
Rolf Carlson, KTH (Sweden)
Björn Granström, KTH (Sweden)

Page (NA) Paper number 361

Abstract:

This paper describes the efforts at KTH in creating educational tools for speech technology. The demand for such tools is increasing with the advent of speech as a medium for man-machine communication. The world wide web was chosen as our platform in order to increase the usability and accessibility of our computer exercises. The aim was to provide dedicated educational software instead of exercises based on complex research tools. Currently, the set of exercises comprises basic speech analysis, multi-modal speech synthesis and spoken dialogue systems. Students access web pages in which the exercises have been embedded as applets. This makes it possible to use them in a classroom setting, as well as from the students? home computers.

SL980361.PDF (From Author) SL980361.PDF (Rasterized)

TOP


Universal Speech Tools: The CSLU Toolkit

Authors:

Stephen Sutton, Center for Spoken Language Understanding, Oregon Graduate Institute (USA)
Ronald A. Cole, Center for Spoken Language Understanding, Oregon Graduate Institute (USA)
Jacques de Villiers, Center for Spoken Language Understanding, Oregon Graduate Institute (USA)
Johan Schalkwyk, Fluent Speech Technologies (USA)
Pieter Vermeulen, Center for Spoken Language Understanding, Oregon Graduate Institute (USA)
Michael W. Macon, Center for Spoken Language Understanding, Oregon Graduate Institute (USA)
Yonghong Yan, Center for Spoken Language Understanding, Oregon Graduate Institute (USA)
Edward Kaiser, Center for Spoken Language Understanding, Oregon Graduate Institute (USA)
Brian Rundle, Center for Spoken Language Understanding, Oregon Graduate Institute (USA)
Khaldoun Shobaki, Center for Spoken Language Understanding, Oregon Graduate Institute (USA)
Paul Hosom, Center for Spoken Language Understanding, Oregon Graduate Institute (USA)
Alex Kain, Center for Spoken Language Understanding, Oregon Graduate Institute (USA)
Johan Wouters, Center for Spoken Language Understanding, Oregon Graduate Institute (USA)
Dominic W. Massaro, University of California, Santa Cruz (USA)
Michael Cohen, University of California, Santa Cruz (USA)

Page (NA) Paper number 649

Abstract:

A set of freely available, universal speech tools is needed to accelerate progress in the speech technology. The CSLU Toolkit represents an effort to make core technology and fundamental infrastructure accessible, affordable and easy to use. The CSLU Toolkit has been under development for five years. This paper describes recent improvements, additions and uses of the CSLU Toolkit.

SL980649.PDF (From Author) SL980649.PDF (Rasterized)

TOP


Creating a Mexican Spanish Version of the CSLU Toolkit

Authors:

Ben Serridge, Universidad de las Americas-Puebla (Mexico)
Alejandro Barbosa, Universidad de las Americas-Puebla (Mexico)
Ronald A. Cole, Center for Spoken Language Understanding, Oregon Graduate Institute of Science and Technology (USA)
Nora Munive, Universidad de las Americas-Puebla (Mexico)
Alcira Vargas, Universidad de las Americas-Puebla (Mexico)

Page (NA) Paper number 923

Abstract:

The CSLU Toolkit is designed to facilitate the rapid development of spoken dialogue systems for a wide variety of applications, as well as to provide a framework for conducting research in the underlying speech technologies. This paper describes the creation of a Mexican Spanish version of the CSLU Toolkit (both synthesis and recognition) undertaken at the Universidad de las Américas in Puebla, México. Based on the Festival Speech Synthesis System of the University of Edinburgh, we have developed a complete concatenative text-to-speech system for Mexican Spanish, which is currently incorporated into the toolkit and includes both a male and female voice. In the area of recognition, we have created a set of task-specific Spanish recognizers for continuous digits, spelled words, and yes/no phrases, as well as a "general-purpose" phonetic recognizer suitable for arbitrary sub-domains. Using the Rapid Application Developer (RAD) component of the CSLU Toolkit, it is now possible to quickly prototype spoken dialogue systems in Spanish. The Spanish components of the CSLU Toolkit are freely available for non-commercial use from the following web page: http://info.pue.udlap.mx/~sistemas/tlatoa.

SL980923.PDF (From Author) SL980923.PDF (Rasterized)

TOP


A Voice User Interface Demonstration System for Mexican Spanish

Authors:

Carmen García-Mateo, E.T.S.I. de Telecomunicación, University of Vigo (Spain)
Qiru Zhou, Dialogue Systems Research Department, Bell Laboratories (USA)
Chin-Hui Lee, Dialogue Systems Research Department, Bell Laboratories (USA)
Andrew Pargellis, Dialogue Systems Research Department, Bell Laboratories (USA)

Page (NA) Paper number 884

Abstract:

We present a Mexican Spanish voice user interface demonstration system. It was built on a speech research platform developed at Bell Labs, which provides major speech technology and interface components, including automatic speech recognition, text-to-speech synthesis, audio input/output functions and telephone interface. The application is written in the PERL script language with an embedded Voice Interface Language (VIL) that connects the speech and interface modules to PERL. Given the set of multilingual speech processing capabilities on the platform and the VIL, we were able to quickly develop a Mexican Spanish system using PERL with speech-enabled messaging and information access functionality similar to our English voice user interface demonstration system.

SL980884.PDF (From Author) SL980884.PDF (Rasterized)

TOP