Punera and
Ghosh extended the idea of hard clustering ensembles to the soft clustering scenario. Each instance in a soft ensemble is represented by a concatenation of
r posterior membership probability distributions obtained from the constituent clustering algorithms. We can define a distance measure between two instances using the
Kullback–Leibler (KL) divergence, which calculates the "distance" between two probability distributions. • '''''': extends CSPA by calculating a similarity matrix. Each object is visualized as a point in dimensional space, with each dimension corresponding to probability of its belonging to a cluster. This technique first transforms the objects into a label-space and then interprets the
dot product between the vectors representing the objects as their similarity. • '''''':extends MCLA by accepting soft clusterings as input. sMCLA's working can be divided into the following steps: • Construct Soft Meta-Graph of Clusters • Group the Clusters into Meta-Clusters • Collapse Meta-Clusters using Weighting • Compete for Objects • ''''
:represents the ensemble as a bipartite graph with clusters and instances as nodes, and edges between the instances and the clusters they belong to. This approach can be trivially adapted to consider soft ensembles since the graph partitioning algorithm METIS accepts weights on the edges of the graph to be partitioned. In sHBGF, the graph has n
+ t'' vertices, where t is the total number of underlying clusters. •
Bayesian consensus clustering (BCC): defines a fully
Bayesian model for soft consensus clustering in which multiple source clusterings, defined by different input data or different probability models, are assumed to adhere loosely to a consensus clustering. The full posterior for the separate clusterings, and the consensus clustering, are inferred simultaneously via
Gibbs sampling. •
Ensemble Clustering Fuzzification Means (ECF-Means): ECF-means is a clustering algorithm, which combines different clustering results in ensemble, achieved by different runs of a chosen algorithm (
k-means), into a single final clustering configuration. == References ==