Bisecting K-Means Clustering on Customer Segments

See Bisecting K-Means Clustering applied to the Customer Segments dataset (200 samples, 5 features). Interactive visualization, metrics, and analysis.

How Bisecting K-Means Clustering Works

Bisecting K-Means recursively splits clusters in half, combining K-Means efficiency with a divisive hierarchical structure.

Bisecting K-Meansdivisive clusteringhierarchicaltop-downsplit history

About the Customer Segments Dataset

200 synthetic customer records with spending, income, and loyalty data. Perfect for market segmentation with clustering.

Samples
200
Features
5
Type
Numeric
Category
Hierarchical

Key Metrics to Watch

Silhouette Score

Measures how similar a point is to its own cluster vs. other clusters. Ranges from −1 to +1; higher is better.

Calinski-Harabasz Index

Ratio of between-cluster to within-cluster variance. Higher values indicate denser, well-separated clusters.

Davies-Bouldin Index

Average similarity between each cluster and its most similar cluster. Lower is better.

Inertia (Within-Cluster SSE)

Sum of squared distances from each point to its assigned centroid. Lower indicates tighter clusters.

When to Use Bisecting K-Means Clustering

Bisecting K-Means Clustering belongs to the Hierarchical family of clustering algorithms. These methods build a tree of clusters, either by merging (agglomerative) or splitting (divisive). They reveal multi-scale structure in data.

Related Examples