Abstract
Probabilistic Decision-Based Neural Networks (PDBNNs) can be considered as a special form of Gaussian Mixture Models (GMMs) with trainable decision thresholds. This paper provides detailed illustrations to compare the recognition accuracy and decision boundaries of PDBNNs with that of GMMs through two pattern recognition tasks, namely the noisy XOR problem and the classification of two-dimensional vowel data. The paper highlights the strengths of PDBNNs by demonstrating that their thresholding mechanism is very effective in detecting data not belonging to any known classes. The original PDBNNs use elliptical basis functions with diagonal covariance matrices, which may be inappropriate for modelling feature vectors with correlated components. This paper overcomes this limitation by using full covariance matrices, and showing that the matrices are effective in characterising non-spherical clusters.
Original language | English |
---|---|
Pages (from-to) | 235-245 |
Number of pages | 11 |
Journal | Neural Computing and Applications |
Volume | 8 |
Issue number | 3 |
DOIs | |
Publication status | Published - 1 Jan 1999 |
Keywords
- EM algorithm
- Gaussian mixture models
- Pattern classification
- Probabilistic decision-based neural networks
ASJC Scopus subject areas
- Software
- Artificial Intelligence