Almost Autonomous Training of Mixtures of Principal Component Analyzers
Loading...

Date
2004
Authors
Journal Title
Journal ISSN
Volume Title
Publisher
Elsevier Science Bv
Open Access Color
Green Open Access
No
OpenAIRE Downloads
OpenAIRE Views
Publicly Funded
No
Abstract
In recent years, a number of mixtures of local PCA models have been proposed. Most of these models require the user to set the number of submodels (local models) in the mixture and the dimensionality of the submodels (i.e., number of PC's) as well. To make the model free of these parameters, we propose a greedy expectation-maximization algorithm to find a suboptimal number of submodels. For a given retained variance ratio, the proposed algorithm estimates for each submodel the dimensionality that retains this given variability ratio. We test the proposed method on two different classification problems: handwritten digit recognition and 2-class ionosphere data classification. The results show that the proposed method has a good performance. (C) 2004 Elsevier B.V. All rights reserved.
Description
Atalay, Volkan/0000-0001-7850-0601
ORCID
Keywords
Pca Mixture Model, Em Algorithm, Regularization
Fields of Science
0301 basic medicine, 03 medical and health sciences, 0202 electrical engineering, electronic engineering, information engineering, 02 engineering and technology
Citation
Musa, MEM; de Ridder, D.; Duin, RPW; Atalay, V., "Almost autonomous training of mixtures of principal component analyzers" Pattern Recognition Letters, Vol.25, No.9, pp.1085-1095, (2004).
WoS Q
Q2
Scopus Q
Q1

OpenCitations Citation Count
6
Source
Pattern Recognition Letters
Volume
25
Issue
9
Start Page
1085
End Page
1095
PlumX Metrics
Citations
CrossRef : 5
Scopus : 5
Captures
Mendeley Readers : 15
SCOPUS™ Citations
5
checked on Feb 23, 2026
Web of Science™ Citations
4
checked on Feb 23, 2026
Page Views
2
checked on Feb 23, 2026
Google Scholar™


