Clustering Methods: K-means, Hierarchical - Higher Difficulty Problems

Q. In hierarchical clustering, what does 'agglomerative' refer to?
  • A. A method that starts with all points as individual clusters
  • B. A method that requires the number of clusters to be predefined
  • C. A technique that merges clusters based on distance
  • D. A type of clustering that uses a centroid
Q. In hierarchical clustering, what is agglomerative clustering?
  • A. A bottom-up approach to cluster formation
  • B. A top-down approach to cluster formation
  • C. A method that requires prior knowledge of clusters
  • D. A technique that uses K-means as a base
Q. In hierarchical clustering, what is the result of a dendrogram?
  • A. A visual representation of the clustering process
  • B. A table of cluster centroids
  • C. A list of data points in each cluster
  • D. A summary of the clustering algorithm's performance
Q. What is a common application of clustering in marketing?
  • A. Predicting customer behavior
  • B. Segmenting customers into distinct groups
  • C. Optimizing supply chain logistics
  • D. Forecasting sales trends
Q. What is a common application of K-means clustering in the real world?
  • A. Image segmentation
  • B. Spam detection
  • C. Sentiment analysis
  • D. Time series forecasting
Q. What is a key advantage of using hierarchical clustering over K-means?
  • A. It requires less computational power
  • B. It does not require the number of clusters to be specified in advance
  • C. It is always more accurate
  • D. It can handle larger datasets
Q. What is a key characteristic of DBSCAN compared to K-means?
  • A. It requires the number of clusters to be specified
  • B. It can find clusters of arbitrary shape
  • C. It is faster than K-means for all datasets
  • D. It uses centroids to define clusters
Q. What is the main advantage of hierarchical clustering over K-means?
  • A. It does not require the number of clusters to be specified in advance
  • B. It is faster and more efficient
  • C. It can handle larger datasets
  • D. It is less sensitive to outliers
Q. What is the main advantage of using hierarchical clustering over K-means?
  • A. It is faster and more efficient
  • B. It does not require the number of clusters to be specified
  • C. It can handle large datasets better
  • D. It is less sensitive to outliers
Q. What is the main difference between K-means and K-medoids clustering?
  • A. K-means uses centroids, while K-medoids uses actual data points
  • B. K-medoids is faster than K-means
  • C. K-means can only handle numerical data, while K-medoids can handle categorical data
  • D. K-medoids requires the number of clusters to be specified, while K-means does not
Q. What is the primary objective of the K-means clustering algorithm?
  • A. To minimize the distance between points in the same cluster
  • B. To maximize the distance between different clusters
  • C. To create a hierarchical structure of clusters
  • D. To classify data into predefined categories
Q. Which distance metric is commonly used in K-means clustering?
  • A. Manhattan distance
  • B. Cosine similarity
  • C. Euclidean distance
  • D. Hamming distance
Q. Which of the following clustering methods is best suited for discovering non-linear relationships in data?
  • A. K-means
  • B. Hierarchical clustering
  • C. DBSCAN
  • D. Gaussian Mixture Models
Q. Which of the following clustering methods is sensitive to outliers?
  • A. K-means
  • B. Hierarchical clustering
  • C. DBSCAN
  • D. Gaussian Mixture Models
Q. Which of the following is NOT a type of hierarchical clustering?
  • A. Single linkage
  • B. Complete linkage
  • C. K-means linkage
  • D. Average linkage
Q. Which of the following scenarios is best suited for hierarchical clustering?
  • A. When the number of clusters is known
  • B. When the data is high-dimensional
  • C. When a hierarchy of clusters is desired
  • D. When speed is a priority
Q. Which of the following scenarios is K-means clustering NOT suitable for?
  • A. When clusters are spherical and evenly sized
  • B. When the number of clusters is known
  • C. When clusters have varying densities
  • D. When outliers are present in the data
Showing 1 to 17 of 17 (1 Pages)