Abstract
Alternative clustering algorithms target finding alternative groupings of a dataset, on which traditional clustering algorithms can find only one even though many alternatives could exist. In this research, we propose a method for finding alternative clusterings of a dataset based on feature selection. Using the finding that each clustering has a set of so-called important features, we find the possible important features for the altenative clustering in subsets of data; we transform the data by weighting these features so that the original clustering will not likely to be found in the new data space. We then use the incremental K-means algorithm to directly maximizes the quality of the new clustering found in the new data space. We compare our approach with some previous works on a collection of machine learning datasets and another collection of documents. Our approach was the most stable one as it resulted in different and high quality clusterings in all of the tests. The results showed that by using feature selection, we can improve the dissimilarity between clusterings, and by directly maximizing the clustering quality, we can also achieve better clustering quality than the other approaches.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Bae, E., Bailey, J.: Coala: A novel approach for the extraction of an alternate clustering of high quality and high dissimilarity. In: ICDM, pp. 53–62 (2006)
Bezdek, J.C., Li, W.Q., Attikiouzel, Y., Windham, M.: A geometric approach to cluster validity for normal mixtures. Soft Computing - A Fusion of Foundations, Methodologies and Applications 1(4), 166–179 (1997)
Dang, X.H., Bailey, J.: Generation of alternative clusterings using the cami approach. In: SDM, pp. 118–129 (2010)
Dang, X.H., Bailey, J.: A hierarchical information theoretic technique for the discovery of non linear alternative clusterings. In: KDD, pp. 573–582. ACM (2010)
Dash, M., Liu, H.: Feature selection for clustering. In: PADKK 2000, pp. 110–121 (2000)
Dhillon, I., Kogan, J., Nicholas, C.: Feature selection and document clustering. In: Survey of Text Mining, pp. 73–100. Springer, Heidelberg (2003)
Gondek, D., Hofmann, T.: Non-redundant data clustering. In: ICDM, pp. 75–82 (2004)
Han, J.: Data Mining: Concepts and Techniques. Morgan Kaufmann Publishers Inc., San Francisco (2005)
Patrikainen, A., Meila, M.: Comparing subspace clusterings. IEEE Transactions on Knowledge and Data Engineering 18(7), 902–916 (2006)
Qi, Z., Davidson, I.: A principled and flexible framework for finding alternative clusterings. In: Elder IV, J.F., Fogelman-Souli, F., Flach, P.A., Zaki, M. (eds.) KDD, pp. 717–726. ACM (2009)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2012 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Tao, V.T., Lee, J. (2012). A Novel Approach for Finding Alternative Clusterings Using Feature Selection. In: Lee, Sg., Peng, Z., Zhou, X., Moon, YS., Unland, R., Yoo, J. (eds) Database Systems for Advanced Applications. DASFAA 2012. Lecture Notes in Computer Science, vol 7238. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-29038-1_35
Download citation
DOI: https://doi.org/10.1007/978-3-642-29038-1_35
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-29037-4
Online ISBN: 978-3-642-29038-1
eBook Packages: Computer ScienceComputer Science (R0)