Skip to main content Accessibility help
×
Home
  • Print publication year: 2014
  • Online publication date: July 2014

6 - Kernel methods for cluster analysis

from Part III - Unsupervised learning models for cluster analysis

Summary

Introduction

The various types of raw data encountered in real-world applications fall into two main categories, vectorial and nonvectorial types. For vectorial data, the Euclidean distance or inner product is often used as the similarity measure of the training vectors: (xi, i = 1,…, N}. This leads to the conventional K-means or SOM clustering methods. This chapter extends these methods to kernel-based cluster discovery and then to nonvectorial clustering applications, such as sequence analysis (e.g. protein sequences and signal motifs) and graph partition problems (e.g. molecular interactions, social networks). The fundamental unsupervised learning theory will be systematically extended to nonvectorial data analysis.

This chapter will cover the following kernel-based unsupervised learning models for cluster discovery.

Section 6.2 explores kernel K-means in intrinsic space. In this basic kernel K-means learning model, the original vectors are first mapped to the basis functions for the intrinsic vector space H, and the mapped vectors will then be partitioned into clusters by the conventional K-means. Because the intrinsic-space approach will not be implementable for some vectorial and all nonvectorial applications, alternative representations need to be pursued. According to Theorem 1.1, the LSP condition holds for K-means. According to Eq. (1.20), this means that the problem formulation may be fully and uniquely characterized by the kernel matrix K associated with the training vector, i.e. without a specific vector space being explicitly defined. In short, the original vector-based clustering criterion is converted to a vector-free clustering criterion.

Related content

Powered by UNSILO