IRMA-International.org: Creator of Knowledge
Information Resources Management Association
Advancing the Concepts & Practices of Information Resources Management in Modern Organizations

K-means Clustering Adopting rbf-Kernel

K-means Clustering Adopting rbf-Kernel
View Sample PDF
Author(s): ABM Shawkat Ali (Central Queensland University, Australia)
Copyright: 2008
Pages: 25
Source title: Data Mining and Knowledge Discovery Technologies
Source Author(s)/Editor(s): David Taniar (Monash University, Australia)
DOI: 10.4018/978-1-59904-960-1.ch006

Purchase

View K-means Clustering Adopting rbf-Kernel on the publisher's website for pricing and purchasing information.

Abstract

Clustering technique in data mining has received a significant amount of attention from machine learning community in the last few years as one of the fundamental research area. Among the vast range of clustering algorithm, K-means is one of the most popular clustering algorithm. In this research we extend K-means algorithm by adding well known radial basis function (rbf) kernel and find better performance than classical K-means algorithm. It is a critical issue for rbf kernel, how can we select a unique parameter for optimum clustering task. This present chapter will provide a statistical based solution on this issue. The best parameter selection is considered on the basis of prior information of the data by Maximum Likelihood (ML) method and Nelder-Mead (N-M) simplex method. A rule based meta-learning approach is then proposed for automatic rbf kernel parameter selection.We consider 112 supervised data set and measure the statistical data characteristics using basic statistics, central tendency measure and entropy based approach. We split this data characteristics using well known decision tree approach to generate the rules. Finally we use the generated rules to select the unique parameter value for rbf kernel and then adopt in K-means algorithm. The experiment has been demonstrated with 112 problems and 10 fold cross validation methods. Finally the proposed algorithm can solve any clustering task very quickly with optimum performance.

Related Content

. © 2023. 34 pages.
. © 2023. 15 pages.
. © 2023. 15 pages.
. © 2023. 18 pages.
. © 2023. 24 pages.
. © 2023. 32 pages.
. © 2023. 21 pages.
Body Bottom