Abstract
Probabilistic linear discriminant analysis (PLDA) has achieved good performance in face recognition and speaker recognition. However, the computation of PLDA using the original formulation is inefficient when there are many training data, especially when the dimensionality of the data is high. Faced with this inefficiency issue, we propose scalable formulations for PLDA. The computation of PLDA using the scalable formulations is more efficient than using the original formulation when dealing with many training data. Using the scalable formulations, the PLDA model can significantly outperform other popular classifiers for speaker recognition, such as Support Vector Machine (SVM) and Gaussian Mixture Model (GMM). Besides of directly using PLDA as a classifier, we may also use PLDA as a feature transformation technique. This PLDA-based feature transformation technique can reduce or expand the original feature dimensionality, and at the same time keep the transformed feature vector approximately following the Gaussian distribution. Our experimental results on speaker recognition and acoustic scene classification demonstrate the effectiveness of applying PLDA for feature transformation. It is then promising to combine PLDA with other classification models for improved performance, extending the utility of PLDA to a wider range of areas.
Original language | English |
---|---|
Article number | 103055 |
Pages (from-to) | 1-15 |
Number of pages | 15 |
Journal | Digital Signal Processing: A Review Journal |
Volume | 114 |
DOIs | |
Publication status | Published - Jul 2021 |
Keywords
- Acoustic signal classification
- Feature transformation
- Probabilistic linear discriminant analysis
- Scalability analysis
ASJC Scopus subject areas
- Signal Processing
- Computer Vision and Pattern Recognition
- Statistics, Probability and Uncertainty
- Computational Theory and Mathematics
- Electrical and Electronic Engineering
- Artificial Intelligence
- Applied Mathematics