Intro to ML: Evaluation (Part 1)

Intro to ML: Evaluation (Part 1)

University

10 Qs

Similar activities

CHAPTER 4: IDS AND IPS

CHAPTER 4: IDS AND IPS

University

10 Qs

MIID2380/IBMS1408 LEC10 2-65 Test

MIID2380/IBMS1408 LEC10 2-65 Test

University

10 Qs

AIML_Day2_(Pandas,Matplol,Regression,ROC)

AIML_Day2_(Pandas,Matplol,Regression,ROC)

University

15 Qs

DP-100 Day 2

DP-100 Day 2

University - Professional Development

10 Qs

Tes Awal AlPro Modul 6

Tes Awal AlPro Modul 6

University

10 Qs

EDA-Quiz No.2 (final term)

EDA-Quiz No.2 (final term)

University

15 Qs

AI Project Cycle

AI Project Cycle

KG - Professional Development

10 Qs

Revision 2 MCA

Revision 2 MCA

University

13 Qs

Intro to ML: Evaluation (Part 1)

Intro to ML: Evaluation (Part 1)

Assessment

Quiz

Created by

Josiah Wang

Computers

University

15 plays

Medium

10 questions

Show all answers

1.

MULTIPLE CHOICE QUESTION

1 min • 1 pt

Our model predicts which people have Covid-19. The model correctly predicts 100 people who have Covid-19, but it misses 50 people who have the illness and falsely predicts that 10 people have the illness when they don't. What is the precision of the model?

91%

66%

Not enough information

Answer explanation

The precision of a model is calculated as the number of true positives (100 in this case) divided by the sum of true positives and false positives (100+10). This gives us 100/110 = 0.909, which is approximately 91%. Therefore, the correct answer is 91%.

2.

MULTIPLE CHOICE QUESTION

1 min • 1 pt

A diagnostic test is used to identify individuals with a certain disease. The test correctly identifies 120 individuals with the disease, but fails to identify 30 individuals who actually have it, and incorrectly identifies 15 individuals as having the disease. What is the recall of the test?

80%

75%

Not enough information

Answer explanation

Recall is calculated as true positives divided by the sum of true positives and false negatives. Here, recall = 120 / (120 + 30) = 120 / 150 = 0.8 or 80%. Thus, the correct answer is 80%.

3.

MULTIPLE CHOICE QUESTION

1 min • 1 pt

If we predict True for every observation, what will our model recall be?

Not enough information

0%

100%

Answer explanation

The recall of a model is calculated as the number of true positives divided by the sum of true positives and false negatives. If we predict True for every observation, there will be no false negatives, hence the recall will be 100%. This is because we are correctly identifying all positive cases, even though we may also be incorrectly identifying some negative cases as positive.

4.

MULTIPLE CHOICE QUESTION

1 min • 1 pt

Which of the following do we use a test set for?

We use it to evaluate new features that we could add to our model

We use it for hyper-parameter tuning

We use it for held-out performance evaluation

None of the above

Answer explanation

The test set is used for held-out performance evaluation. This means it is used to assess the performance of a model on unseen data, providing an unbiased evaluation of the final model fit. It is not used for hyper-parameter tuning or evaluating new features for the model.

5.

MULTIPLE CHOICE QUESTION

1 min • 1 pt

When is cross validation particularly helpful?

When we have a small dataset

When we have a large dataset

When we are particularly constrained on computational resources

Answer explanation

Cross validation is particularly helpful when we have a small dataset because it allows us to maximize the use of limited data for both training and validation. By dividing the dataset into multiple folds and iteratively training and validating on different subsets, we can obtain a more accurate estimate of the model's performance and reduce the risk of overfitting.

6.

MULTIPLE CHOICE QUESTION

1 min • 1 pt

Which of the following does an F1 score depend on?

TP, FP, FN

TP, TN, FP, TN

TN, FP, TP

TP, FN

Answer explanation

The F1 score is a measure of a test's accuracy and it depends on True Positive (TP), False Positive (FP), and False Negative (FN). It is the harmonic mean of precision and recall, where precision is TP/(TP+FP) and recall is TP/(TP+FN). Therefore, the correct answer is 'TP, FP, FN'.

7.

MULTIPLE CHOICE QUESTION

1 min • 1 pt

Which of the following could you use to evaluate performance on a regression task?

Precision

Recall

Accuracy

MSE

Answer explanation

MSE, or Mean Squared Error, is the correct choice for evaluating performance on a regression task. It measures the average of the squares of the errors—that is, the average squared difference between the estimated values and the actual value. Precision, Recall, and Accuracy are metrics used for classification tasks, not regression tasks.

8.

MULTIPLE CHOICE QUESTION

1 min • 1 pt

What is the accuracy of a model that correctly predicts 150 cases out of 200 total cases, with 30 false positives?

75%

85%

Not enough information

Answer explanation

Accuracy is calculated as (True Positives + True Negatives) / Total Cases. Here, (True Positives + True Negatives) = 150, Total Cases = 200. Thus, Accuracy = 150/200 = 0.75 or 75%. Therefore, the correct answer is 75%.

9.

MULTIPLE CHOICE QUESTION

1 min • 1 pt

Which metric is most appropriate for evaluating a model when the cost of false negatives is very high?

Precision

Recall

F1 Score

Accuracy

Answer explanation

Recall is the most appropriate metric when false negatives are costly, as it measures the ability of the model to identify all relevant instances. High recall ensures that most actual positives are correctly predicted.

10.

MULTIPLE CHOICE QUESTION

1 min • 1 pt

In a classification task, if a model has a high precision but low recall, what does this indicate?

The model is good at identifying all positive cases

The model is good at identifying positive cases but misses many

The model is not good at identifying positive cases

The model has a balanced performance

Answer explanation

High precision means the model correctly identifies many positive cases, but low recall indicates it misses a significant number of them. Thus, the model is good at identifying positive cases but misses many.

Explore all questions with a free account

or continue with
Microsoft
Apple
Others
By signing up, you agree to our Terms of Service & Privacy Policy
Already have an account?