Abstract
We explore a method to boost discriminative capabilities of Probabilistic Linear Discriminant Analysis (PLDA) model without losing its generative advantages. To this end, our focus is in a low-dimensional PLDA latent subspace. We optimize the model with respect to MMI (Maximum Mutual Information) and our own objective functions, which is an approximation to the detection cost function. We evaluate the performance on NIST Language Recognition Evaluation 2015. Our model trains faster and performs more accurately in comparison to both generative PLDA and discriminative LDA baselines with 12% and 4% relative improvement in the average detection cost, respectively. The proposed method is applicable for a broad range of closed-set tasks.
Original language | English |
---|---|
Pages | 81-88 |
Number of pages | 8 |
DOIs | |
Publication status | Published - Jun 2016 |
Externally published | Yes |
Event | Speaker and Language Recognition Workshop, Odyssey 2016 - Bilbao, Spain Duration: 21 Jun 2016 → 24 Jun 2016 |
Conference
Conference | Speaker and Language Recognition Workshop, Odyssey 2016 |
---|---|
Country/Territory | Spain |
City | Bilbao |
Period | 21/06/16 → 24/06/16 |
ASJC Scopus subject areas
- Signal Processing
- Software
- Human-Computer Interaction