0% found this document useful (0 votes)
246 views3 pages

Nptel Week 8

The document discusses decision trees and related machine learning concepts like bagging, boosting, precision, recall, and overfitting. It provides 15 multiple choice questions about decision tree algorithms, their applications, techniques like pruning and encoding to improve performance, and metrics like precision.

Uploaded by

deepankarweb2
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
246 views3 pages

Nptel Week 8

The document discusses decision trees and related machine learning concepts like bagging, boosting, precision, recall, and overfitting. It provides 15 multiple choice questions about decision tree algorithms, their applications, techniques like pruning and encoding to improve performance, and metrics like precision.

Uploaded by

deepankarweb2
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 3

BUSINESS INTELLIGENCE AND ANALYTICS

ASSIGNMENT WEEK 8:

Total marks = 1 Marks


15 Qns * 1 marks = 15 marks

1. Which of the following is a common method for splitting nodes in a decision tree? (1
Mark)

A. Gini impurity
B. Cross-validation
C. Gradient descent
D. Principal component analysis
Answer: A

2. What is the purpose of pruning in decision trees? (1 Mark)

A. To reduce the depth of the tree and prevent overfitting


B. To optimize the tree's parameters
C. To handle missing data
D. To reduce the depth of the tree and prevent underfitting

Answer: A

3. Which of the following is a popular algorithm for constructing decision trees? (1 Mark)

A. ID3
B. k-Nearest Neighbors
C. Support Vector Machines
D. Naive Bayes
Answer: A

4. What is the main difference between classification and regression trees in a CART
algorithm? (1 Mark)
A. Classification trees predict categorical variables, while regression trees predict
continuous variables
B. Classification trees use Gini impurity as the splitting criterion, while regression trees
use information gain
C. Classification trees can handle missing data, while regression trees cannot
D. Classification trees are computationally expensive, while regression trees are
computationally inexpensive
Answer: A

5. Consider the following statements:


a) The term Boosting stands for bootstrap aggregation.
b) Bagging is less susceptible to model overfitting as compared to Boosting
Which of the above statements are correct? (1 Mark)
A. Only a
B. Only b
C. Both a and b
D. Neither a nor b

Ans: B. Only b

6. What is entropy in the context of decision trees? (1 Mark)

A. A measure of disorder or impurity in a node


B. A measure of the complexity of a decision tree
C. The difference between the predicted and actual values in a node
D. The rate at which information is gained in a decision tree
Answers: A

7. Which of the following is a common stopping criterion for growing a decision tree? (1
Mark)

A. Reaching a maximum depth


B. Achieving a minimum information gain
C. Achieving a minimum Gini impurity
D. Both A and B
Answer: D

8. For decision trees, what purpose does "one-hot encoding" serve? (1 Mark)
A. Handling missing data
B. Transform categorical data into numerical format that algorithms can process
C. Normalizing continuous variables
D. Reducing the dimensionality of the feature space

Answer: B

9. What's the primary drawback of utilizing a substantial maximum depth for a decision
tree? (1 Mark)
A. It leads to overfitting
B. It cannot capture the noise in the training data
C. It simplifies the computational complexity of the tree
D. It results in the tree underfitting the data

Answer: A

10. Which strategy is effective in mitigating overfitting in decision trees? (1 Mark)


A. Pruning
B. Bagging
C. Boosting
D. All of the above

Answer: D

11. Which of the following is NOT commonly associated with the use of decision trees? (1
Mark)

A. Fraud detection
B. Stock price prediction

C. Customer churn prediction

D. Image classification

Answer: D

12. How can decision trees be made more robust to noise in the data? (1 Mark)

A. By increasing the maximum depth of the tree

B. By using a smaller minimum samples per leaf

C. By using ensemble techniques like bagging or boosting

D. By removing features with low importance

Answer: C

13. What role do leaf nodes play in a decision tree? (1 Mark)

A. To hold the conditions for data splitting

B. To represent the importance of a feature

C. To signify the depth of the tree

D. To denote the class label or predicted value

Answer: D

14. If the true positive value is 10 and the false positive value is 15, what is the precision score
for the classification model? (1 Mark)

A. 0.6

B. 0.4

C. 0.5

D. None of the above

Answer: B

15. Which of the following definitions describes false negatives? (1 Mark)


A. Predicted negatives that are actually positives
B. Predicted positives that are actually negatives
C. Predicted negatives that are actually negatives
D. Predicted positives that are actually positives

Answer: A

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy