Spacer ICASSP '98 Main Page

Spacer
General Information
Spacer
Conference Schedule
Spacer
Technical Program
Spacer
    Overview
    50th Annivary Events
    Plenary Sessions
    Special Sessions
    Tutorials
    Technical Sessions
    
By Date
    May 12, Tue
May 13, Wed
May 14, Thur
May 15, Fri
    
By Category
    AE    ANNIV   
COMM    DSP   
IMDSP    MMSP   
NNSP    PLEN   
SP    SPEC   
SSAP    UA   
VLSI   
    
By Author
    A    B    C    D    E   
F    G    H    I    J   
K    L    M    N    O   
P    Q    R    S    T   
U    V    W    X    Y   
Z   

    Invited Speakers
Spacer
Registration
Spacer
Exhibits
Spacer
Social Events
Spacer
Coming to Seattle
Spacer
Satellite Events
Spacer
Call for Papers/
Author's Kit

Spacer
Future Conferences
Spacer
Help

Abstract -  NNSP5   


 
NNSP5.1

   
Minimum Detection Error Training for Acoustic Signal Monitoring
H. Watanabe, Y. Matsumoto, S. Katagiri  (ATR Human Information Processing Labs, Japan)
In this paper we propose a novel approach to the detection of acoustic irregular signals using Minimum Detection Error (MDE) training. The MDE training is based on the Generalized Probabilistic Descent for discriminative pattern recognizer design. We demonstrate its fundamental utility by experiments in which several acoustic events are detected in a noisy environment.
 
NNSP5.2

   
Fast Subspace Tracking and Neural Network Learning by a Novel Information Criterion
Y. Miao, Y. Hua  (University of Melbourne, Australia)
We introduce a novel information criterion (NIC) for searching for the optimum weights of a two-layer linear neural network (NN). The NIC exhibits a single global maximum attained if and only if the weights span the (desired) principal subspace of a covariance matrix. The other stationary points of the NIC are (unstable) saddle points. We develop an adaptive algorithm based on the NIC for estimating and tracking the principal subspace of a vector sequence. The NIC algorithm provides a fast on-line learning of the optimum weights for the two-layer linear NN. The NIC algorithm has several key advantages such as faster convergence which is illustrated through analysis and simulation.
 
NNSP5.3

   
Adaptive Regularization of Neural Networks Using Conjugate Gradient
C. Goutte, J. Larsen  (Technical University of Denmark, Denmark)
Recently we suggested a regularization scheme which iteratively adapts regularization parameters by minimizing validation error using simple gradient descent. In this contribution we present an improved algorithm based on the conjugate gradient technique. Numerical experiments with feed-forward neural networks successfully demonstrate improved generalization ability and lower computational cost.
 
NNSP5.4

   
Design of Robust Neural Network Classifiers
J. Larsen, L. Andersen, M. Hintz-Madsen, L. Hansen  (Technical University of Denmark, Denmark)
This paper addresses a new framework for designing robust neural network classifiers. The network is optimized using the maximum a posteriori technique, i.e., the cost function is the sum of the log-likelihood and a regularization term (prior). In order to perform robust classification, we present a modified likelihood function which incorporate the potential risk of outliers in the data. This leads to introduction of a new parameter, the outlier probability. Designing the neural classifier involves optimization of network weights as well as outlier probability and regularization parameters. We suggest to adapt the outlier probability and regularization parameters by minimizing the error on a validation set, and a simple gradient descent scheme is derived. In addition, the framework allows for constructing a simple outlier detector. Experiments with artificial data demonstrates the potential of the suggested framework.
 
NNSP5.5

   
Extraction of Independent Components from Hybrid Mixture: KuicNet Learning Algorithm and Applications
S. Kung  (Princeton University, USA);   C. Mejuto  (University of La Coruna, Spain)
A hybrid mixture is a mixture of supergaussian, gaussian, and subgaussian independent components(ICs). This paper addresses extraction of ICs from a hybrid mixture. There are two (single-output vs. all-outputs) approaches to the design of contrast functions. We advocate the former approach due to its (1) simple and closed-form analysis, and (2) numerical convergence andcomputational saving. Via this approach, the positive kurtosis (resp. negative kurtosis) can be proved to be a valid contrast function for extracting supergaussian (resp. subgaussian) ICs from any nontrivial hybrid mixture. We shall also develop a network algorithm, Ku}rtosis-based {\ul I}ndependent {\ul C}omponent {\ul Net}work(KuicNet), for recursively extracting ICs. Numerical and convergence properties are analyzed and several application examples demonstrated.
 
NNSP5.6

   
Why Natural Gradient?
S. Amari  (RIKEN Brain Science Institute, Japan);   S. Douglas  (University of Utah, USA)
Gradient adaptation is a useful technique for adjusting a set of parameters to minimize a cost function. While often easy to implement, the convergence speed of gradient adaptation can be slow when the slope of the cost function varies widely for small changes in the parameters. In this paper, we outline an alternative technique, termed natural gradient adaptation, that overcomes the poor convergence properties of gradient adaptation in many cases. The natural gradient is based on differential geometry and employs knowledge of the Riemannian structure of the parameter space to adjust the gradient search direction. Unlike Newton's method, natural gradient adaptation does not assume a locally-quadratic cost function. Moreover, for maximum likelihood estimation tasks, natural gradient adaptation is asymptotically Fisher-efficient. A simple example illustrates the desirable properties of natural gradient adaptation.
 
NNSP5.7

   
Neural Network Inversion of Snow Parameters by Fusion of Snow Hydrology Prediction and SSM/I Microwave Satellite Measurements
Y. Wang, J. Hwang, C. Chen, L. Tsang, B. Nijssen, D. Lettenmaier  (University of Washington, USA)
Inverse remote sensing problems are generally ill-posed. In this paper, we propose an approach, which integrates the dense media radiative transfer (DMRT) model, snow hydrology model, neural networks and SSM/I microwave measurements, to infer the snow depth. Four multilayer perceptrons (MLPs) were trained using the data from DMRT model. With the provision of initial guess from snow hydrology prediction, neural networks effectively invert the snow parameters based on SSM/I measurements. In addition, a prediction neural network is used to achieve adaptive learning rates and good initial estimate of snow depth for inversion. Result shows that our algorithm can effectively and accurately retrieve snow parameters from these highly nonlinear and many-to-one mappings.
 
NNSP5.8

   
A Piecewise Linear Recurrent Neural Network Structure and its Dynamics
X. Liu  (GlobeSpan Technologies Inc., USA);   T. Adali, L. Demirekler  (University of Maryland, Baltimore, USA)
We present a piecewise linear recurrent neural network (PL-RNN) structure by combining the canonical piecewise linear function with the autoregressive moving average (ARMA) model such that an augmented input space is partitioned into regions where an ARMA model is used in each. The piecewise linear structure allows for easy implementation, and in training, allows for use of standard linear adaptive filtering techniques based on gradient optimization and description of convergence regions for the step-size. We study the dynamics of PL-RNN and show that it defines a contractive mapping and is bounded input bounded output stable. We introduce application of PL-RNN to channel equalization and show that it closely approximates the performance of the traditional RNN that uses sigmoidal activation functions.
 

< Previous Abstract - NNSP4