Q: A junior data analyst uses tree-based learning for a sales and marketing project. Currently, they are interested in the section of the tree that represents where the first decision is made. What are they examining?
- Branches
- Leaves
- Roots
- Splits
Q: What are some disadvantages of decision trees? Select all that apply.
- Preparing data to train a decision is a complex process involving significant preprocessing
- Decision trees require assumptions regarding the distribution of underlying data.
- Decision trees can be particularly susceptible to overfitting.
- When new data is introduced, decision trees can be less effective at prediction.
Q: Which section of a decision tree is where the final prediction is
made?
- Decision node
- Split
- Leaf node
- Root node
Q: In a decision tree ensemble model, which hyperparameter controls how
many decision trees the model will build for its ensemble?
- max_features
- max_depth
- n_trees
- n_estimators
Q: What process uses different “folds” (portions) of the data to train
and evaluate a model across several iterations?
- Grid search
- Model validation
- Cross validation
- Proportional verification
Q: Which of the following statements correctly describe ensemble
learning? Select all that apply.
- When building an ensemble using different types of models, each should be trained on completely different data.
- Predictions using an ensemble of models can be accurate even when the individual models are barely more accurate than a random guess.
- Ensemble learning involves aggregating the outputs of multiple models to make a final prediction.
- If a base learner’s prediction is only slightly better than a random guess, it is called a “weak learner.”
Q: Fill in the blank: A random forest is an ensemble of decision-tree
_____ that are trained on bootstrapped data.
- Statements
- Observations
- base learners
- variables
Q: What are some benefits of boosting? Select all that apply.
- Boosting is the most interpretable model methodology.
- Boosting is a powerful predictive methodology.
- Boosting can handle both numeric and categorical features.
- Boosting does not require the data to be scaled.
Q: Which of the following statements correctly describe gradient
boosting? Select all that apply.
- Gradient boosting machines cannot perform classification tasks.
- Gradient boosting machines have many hyperparameters.
- Gradient boosting machines do not give coefficients or directionality for their individual features.
- Gradient boosting machines are often called black-box models because their predictions can be difficult to explain.
Q: A data professional uses tree-based learning for an operations
project. Currently, they are interested in the nodes at which the trees split.
What type of nodes do they examine?
- Decision
- Branch
- Leaf
- Root
Q: What are some benefits of decision trees? Select all that apply.
- When working with decision trees, overfitting is unlikely.
- When preparing data to train a decision tree, very little preprocessing is required.
- Decision trees enable data professionals to make predictions about future events based on currently available information.
- Decision trees require no assumptions regarding the distribution of underlying data.
Q: In a decision tree, what type(s) of nodes can decision nodes point
to? Select all that apply.
- Split
- Root node
- Leaf node
- Decision node
Q: In a decision tree model, which hyperparameter sets the threshold
below which nodes become leaves?
- Min child weight
- Min samples tree
- Min samples split
- Min samples leaf
Q: When might you use a separate validation dataset? Select all that
apply.
- If you have very little data.
- If you want to choose the specific samples used to validate the model.
- If you have a very large amount of data.
- If you want to compare different model scores to choose a champion before predicting on test holdout data.
Q: What tool is used to confirm that a model achieves its intended
purpose by systematically checking combinations of hyperparameters to identify
which set produces the best results, based on the selected metric?
- GridSearchCV
- Model validation
- Cross validation
- Hyperparameter verification
Q: Which of the following statements correctly describe ensemble
learning? Select all that apply.
- If a base learner’s prediction is equally effective as a random guess, it is a strong learner.
- It’s possible to use the same methodology for each contributing model, as long as there are numerous base learners.
- Ensemble learning involves building multiple models.
- It’s possible to use very different methodologies for each contributing model.
Q: Which of the following statements correctly describe gradient
boosting? Select all that apply.
- Gradient boosting machines build models in parallel.
- Gradient boosting machines tell you the coefficients for each feature.
- Gradient boosting machines work well with missing data.
- Gradient boosting machines do not require the data to be scaled.
Q: Which of the following statements accurately describe decision
trees? Select all that apply.
- Decision trees are equally effective at predicting both existing and new data.
- Decision trees work by sorting data.
- Decision trees require no assumptions regarding the distribution of underlying data.
- Decision trees are susceptible to overfitting.
Q: What is the only section of a decision tree that contains no
predecessors?
- Leaf node
- Root node
- Decision node
- Split based on what will provide the most predictive power.
Q: In a decision tree, nodes are where decisions are made, and they are
connected by edges.
- True
- False