Hierarchical Clustering Basics Quiz

Reviewed by Editorial Team
The ProProfs editorial team is comprised of experienced subject matter experts. They've collectively created over 10,000 quizzes and lessons, serving over 100 million users. Our team includes in-house content moderators and subject matter experts, as well as a global network of rigorously trained contributors. All adhere to our comprehensive editorial guidelines, ensuring the delivery of high-quality content.
Learn about Our Editorial Process
| By Thames
T
Thames
Community Contributor
Quizzes Created: 6575 | Total Attempts: 67,424
| Questions: 15 | Updated: May 2, 2026
Please wait...
Question 1 / 16
🏆 Rank #--
0 %
0/100
Score 0/100

1. What is hierarchical clustering?

Explanation

Hierarchical clustering is an unsupervised learning technique that organizes data into a tree-like structure, known as a dendrogram. It works by either merging similar data points into larger clusters or splitting them into smaller clusters, allowing for a comprehensive view of the data's structure and relationships at various levels of granularity.

Submit
Please wait...
About This Quiz
Hierarchical Clustering Basics Quiz - Quiz

Test your understanding of hierarchical clustering, a fundamental unsupervised learning technique. This Hierarchical Clustering Basics Quiz covers dendrograms, linkage methods, distance metrics, and practical applications. Ideal for college students mastering clustering algorithms and their real-world use cases.

2.

What first name or nickname would you like us to use?

You may optionally provide this to label your report, leaderboard, or certificate.

2. Which of the following best describes a dendrogram?

Explanation

A dendrogram visually represents the arrangement of clusters in hierarchical clustering. It illustrates how clusters are formed by merging data points or groups, showing their relationships and the sequence of merges. This makes it easier to understand the structure and similarity between different clusters within the dataset.

Submit

3. What are the two main approaches to hierarchical clustering?

Explanation

Hierarchical clustering consists of two primary methods: agglomerative and divisive. Agglomerative clustering starts with individual data points and merges them into clusters, while divisive clustering begins with one large cluster and recursively splits it into smaller clusters. These approaches allow for the creation of a hierarchy of clusters, facilitating better data organization and analysis.

Submit

4. In agglomerative clustering, which step occurs first?

Explanation

In agglomerative clustering, the initial step involves treating each data point as its own individual cluster. This sets the foundation for the algorithm, which then iteratively merges the closest clusters until the desired number of clusters is achieved. This approach allows for a systematic buildup of clusters based on proximity.

Submit

5. Which linkage method uses the minimum distance between clusters?

Explanation

Single linkage, also known as nearest neighbor linkage, measures the distance between the closest points of two clusters. This method focuses on the minimum distance, allowing it to effectively identify and merge clusters that are close together, which can lead to elongated cluster shapes in hierarchical clustering.

Submit

6. Complete linkage considers the ____ distance between two clusters.

Explanation

Complete linkage clustering measures the distance between two clusters by considering the maximum distance between any single pair of points from each cluster. This approach ensures that the most dissimilar points are taken into account, leading to a more conservative merging of clusters and maintaining compactness within them.

Submit

7. Ward linkage minimizes the increase in within-cluster variance when merging clusters.

Explanation

Ward linkage is a hierarchical clustering method that aims to minimize the total within-cluster variance. By merging clusters that result in the smallest increase in this variance, it ensures that the resulting clusters are as compact and distinct as possible, leading to more meaningful groupings in the data.

Submit

8. Which distance metric is most commonly used in hierarchical clustering?

Explanation

Euclidean distance is the most commonly used metric in hierarchical clustering because it effectively measures the straight-line distance between two points in Euclidean space. This metric captures the geometric relationships between data points, making it suitable for clustering tasks where the shape and distribution of the data matter.

Submit

9. In divisive hierarchical clustering, you start with ____ cluster(s).

Explanation

In divisive hierarchical clustering, the process begins with a single cluster that encompasses all data points. This cluster is then recursively split into smaller clusters based on dissimilarity until each data point forms its own cluster or a stopping criterion is met. This approach contrasts with agglomerative clustering, where individual data points start as separate clusters.

Submit

10. What is the primary advantage of hierarchical clustering over k-means?

Explanation

Hierarchical clustering offers the significant advantage of not needing a predetermined number of clusters, allowing for more flexibility in data analysis. This feature enables users to explore the data structure at various levels of granularity, making it easier to identify natural groupings without forcing the data into a specific number of clusters.

Submit

11. A dendrogram's horizontal cut line determines the final ____ of clusters.

Explanation

A dendrogram visually represents hierarchical clustering, where the horizontal cut line signifies the threshold for merging clusters. The position of this line indicates how many clusters will remain after the cut, determining the final number of clusters formed from the data based on the specified similarity or dissimilarity.

Submit

12. Average linkage computes the distance between two clusters as the average of all pairwise distances.

Explanation

Average linkage is a hierarchical clustering method that determines the distance between two clusters by calculating the average of the distances between all possible pairs of points, one from each cluster. This approach takes into account the overall distribution of points in the clusters, providing a more balanced measure of similarity than single or complete linkage methods.

Submit

13. Which of these is a disadvantage of hierarchical clustering?

Submit

14. The cophenetic correlation coefficient measures how well a dendrogram preserves the original distances between data points.

Submit

15. In hierarchical clustering, which linkage method tends to produce elongated, chain-like clusters?

Submit
×
Saved
Thank you for your feedback!
View My Results
Cancel
  • All
    All (15)
  • Unanswered
    Unanswered ()
  • Answered
    Answered ()
What is hierarchical clustering?
Which of the following best describes a dendrogram?
What are the two main approaches to hierarchical clustering?
In agglomerative clustering, which step occurs first?
Which linkage method uses the minimum distance between clusters?
Complete linkage considers the ____ distance between two clusters.
Ward linkage minimizes the increase in within-cluster variance when...
Which distance metric is most commonly used in hierarchical...
In divisive hierarchical clustering, you start with ____ cluster(s).
What is the primary advantage of hierarchical clustering over k-means?
A dendrogram's horizontal cut line determines the final ____ of...
Average linkage computes the distance between two clusters as the...
Which of these is a disadvantage of hierarchical clustering?
The cophenetic correlation coefficient measures how well a dendrogram...
In hierarchical clustering, which linkage method tends to produce...
play-Mute sad happy unanswered_answer up-hover down-hover success oval cancel Check box square blue
Alert!