Decision Trees and Random Forests - Higher Difficulty Problems

Download Q&A
Q. In a Decision Tree, what does the Gini impurity measure?
  • A. The accuracy of the model.
  • B. The likelihood of misclassifying a randomly chosen element.
  • C. The depth of the tree.
  • D. The number of features used.
Q. In Random Forests, how are the individual trees trained?
  • A. On the entire dataset without any modifications.
  • B. Using a bootstrapped sample of the dataset.
  • C. On a subset of features only.
  • D. Using the same random seed for all trees.
Q. In Random Forests, what does 'bagging' refer to?
  • A. Using all available features for each tree.
  • B. Randomly selecting subsets of data to train each tree.
  • C. Combining predictions from multiple models.
  • D. Pruning trees to improve performance.
Q. In the context of Decision Trees, what does 'feature importance' refer to?
  • A. The number of times a feature is used in the tree.
  • B. The contribution of a feature to the model's predictions.
  • C. The correlation of a feature with the target variable.
  • D. The depth of a feature in the tree.
Q. What is a potential drawback of using a very deep Decision Tree?
  • A. It may not capture complex patterns.
  • B. It can lead to overfitting.
  • C. It requires more computational resources.
  • D. It is less interpretable.
Q. What is the effect of increasing the number of trees in a Random Forest?
  • A. It always increases the training time.
  • B. It can improve model accuracy but may lead to diminishing returns.
  • C. It decreases the model's interpretability.
  • D. It reduces the model's variance but increases bias.
Q. What is the primary advantage of using Random Forests over a single Decision Tree?
  • A. Random Forests are easier to interpret.
  • B. Random Forests reduce overfitting by averaging multiple trees.
  • C. Random Forests require less computational power.
  • D. Random Forests can only handle categorical data.
Q. What is the primary purpose of using ensemble methods like Random Forests?
  • A. To simplify the model.
  • B. To improve prediction accuracy by combining multiple models.
  • C. To reduce the training time.
  • D. To increase interpretability.
Q. What is the purpose of the 'min_samples_split' parameter in a Decision Tree?
  • A. To control the minimum number of samples required to split an internal node.
  • B. To set the maximum depth of the tree.
  • C. To determine the minimum number of samples in a leaf node.
  • D. To specify the maximum number of features to consider.
Q. What is the role of the 'max_features' parameter in a Random Forest model?
  • A. It determines the maximum number of trees in the forest.
  • B. It specifies the maximum number of features to consider when looking for the best split.
  • C. It sets the maximum depth of each tree.
  • D. It controls the minimum number of samples required to split an internal node.
Q. Which evaluation metric is most appropriate for assessing the performance of a Decision Tree on an imbalanced dataset?
  • A. Accuracy
  • B. F1 Score
  • C. Mean Squared Error
  • D. R-squared
Q. Which of the following is a common method for preventing overfitting in Decision Trees?
  • A. Increasing the maximum depth of the tree.
  • B. Pruning the tree after it has been fully grown.
  • C. Using more features.
  • D. Decreasing the number of samples.
Q. Which of the following statements about Decision Trees is true?
  • A. They can only be used for classification tasks.
  • B. They are sensitive to small changes in the data.
  • C. They require feature scaling.
  • D. They cannot handle missing values.
Showing 1 to 13 of 13 (1 Pages)
Soulshift Feedback ×

On a scale of 0–10, how likely are you to recommend The Soulshift Academy?

Not likely Very likely