Back to Search
Start Over
Maximum Joint Probability With Multiple Representations for Clustering.
- Source :
- IEEE Transactions on Neural Networks & Learning Systems; Sep2022, Vol. 33 Issue 9, p4300-4310, 11p
- Publication Year :
- 2022
-
Abstract
- Classical generative models in unsupervised learning intend to maximize $p(X)$. In practice, samples may have multiple representations caused by various transformations, measurements, and so on. Therefore, it is crucial to integrate information from different representations, and lots of models have been developed. However, most of them fail to incorporate the prior information about data distribution $p(X)$ to distinguish representations. In this article, we propose a novel clustering framework that attempts to maximize the joint probability of data and parameters. Under this framework, the prior distribution can be employed to measure the rationality of diverse representations. $K$ -means is a special case of the proposed framework. Meanwhile, a specific clustering model considering both multiple kernels and multiple views is derived to verify the validity of the designed framework and model. [ABSTRACT FROM AUTHOR]
- Subjects :
- PROBABILITY theory
PROBABILISTIC generative models
DATA distribution
Subjects
Details
- Language :
- English
- ISSN :
- 2162237X
- Volume :
- 33
- Issue :
- 9
- Database :
- Complementary Index
- Journal :
- IEEE Transactions on Neural Networks & Learning Systems
- Publication Type :
- Periodical
- Accession number :
- 158869779
- Full Text :
- https://doi.org/10.1109/TNNLS.2021.3056420