How Agglomerative Hierarchical Clustering Works
Agglomerative clustering builds a hierarchy bottom-up, merging the closest pairs of clusters at each step. Dendrogram visualization reveals cluster relationships.
About the Customer Segments Dataset
200 synthetic customer records with spending, income, and loyalty data. Perfect for market segmentation with clustering.
- Samples
- 200
- Features
- 5
- Type
- Numeric
- Category
- Hierarchical
Key Metrics to Watch
Silhouette Score
Measures how similar a point is to its own cluster vs. other clusters. Ranges from −1 to +1; higher is better.
Calinski-Harabasz Index
Ratio of between-cluster to within-cluster variance. Higher values indicate denser, well-separated clusters.
Davies-Bouldin Index
Average similarity between each cluster and its most similar cluster. Lower is better.
Inertia (Within-Cluster SSE)
Sum of squared distances from each point to its assigned centroid. Lower indicates tighter clusters.
When to Use Agglomerative Hierarchical Clustering
Agglomerative Hierarchical Clustering belongs to the Hierarchical family of clustering algorithms. These methods build a tree of clusters, either by merging (agglomerative) or splitting (divisive). They reveal multi-scale structure in data.
Related Examples
K-Means Clustering on Customer Segments
See K-Means Clustering applied to the Customer Segments dataset (200 samples, 5 features). Interactive visualization, metrics, and analysis.
K-Medoids Clustering on Customer Segments
See K-Medoids Clustering applied to the Customer Segments dataset (200 samples, 5 features). Interactive visualization, metrics, and analysis.
DBSCAN Clustering on Customer Segments
See DBSCAN Clustering applied to the Customer Segments dataset (200 samples, 5 features). Interactive visualization, metrics, and analysis.
HDBSCAN Clustering on Customer Segments
See HDBSCAN Clustering applied to the Customer Segments dataset (200 samples, 5 features). Interactive visualization, metrics, and analysis.