Abstract
High-dimensional matrix-valued data is common in scientific and engineering studies and its classification is a significant topic in current statistics. In practice, the discriminative signals of the matrix covariates are oftentimes low rank and sparse. Motivated by this, we propose a sparse and reduced-rank matrix linear discriminant analysis called 'Sr-LDA' for binary classification of high-dimensional matrix-valued data. Specifically, based on the Bayes' linear discriminant rule, we derive the theoretically optimal discriminative matrix-valued covariates under the matrix normal assumptions, and constructed a convex empirical loss function for the estimation of the optimal discriminative matrix-valued covariates under the ℓ 1-norm and nuclear norm penalties. Finite sample error bounds for parameter estimation and the misclassification rate are established. The superior performance of the proposed Sr-LDA is illustrated via extensive simulation and real data studies with comparison to other state-of-The-Art classifiers.
| Original language | English |
|---|---|
| Article number | 3378578 |
| Pages (from-to) | 1134-1138 |
| Number of pages | 5 |
| Journal | IEEE Signal Processing Letters |
| Volume | 31 |
| DOIs | |
| Publication status | Published - Apr 2024 |
Keywords
- classification
- low rank
- Matrix-valued data
- nuclear norm
- sparsity
- ℓ-norm
ASJC Scopus subject areas
- Signal Processing
- Electrical and Electronic Engineering
- Applied Mathematics