Intro to ML: Neural Networks Lecture 2 Part 1

Intro to ML: Neural Networks Lecture 2 Part 1

University

6 Qs

Similar activities

Strong and Weak Residual Plots

Strong and Weak Residual Plots

11th Grade - University

11 Qs

Quiz 12

Quiz 12

University

10 Qs

Scatterplots and Regression

Scatterplots and Regression

9th Grade - University

10 Qs

Data Mining

Data Mining

University

7 Qs

Introduction to Deep Learning

Introduction to Deep Learning

University

10 Qs

AP Statistics Unit 3 Review

AP Statistics Unit 3 Review

12th Grade - University

10 Qs

MIID2380/IBMS1408 LEC10 2-65 Test

MIID2380/IBMS1408 LEC10 2-65 Test

University

10 Qs

AP Statistics Unit 9

AP Statistics Unit 9

12th Grade - University

10 Qs

Intro to ML: Neural Networks Lecture 2 Part 1

Intro to ML: Neural Networks Lecture 2 Part 1

Assessment

Quiz

Created by

Josiah Wang

Mathematics, Computers, Fun

University

42 plays

Hard

6 questions

Show all answers

1.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

Mean squared error is a common loss function for which task?

Regression

Classification

None of these

Regression and Classification

Answer explanation

Mean squared error is the expected (squared) distance between the models predictions and the true values. It therefore is well suited to regression tasks where the label space is a continuous one. MSE assumes the data is normally distributed however, in binary classification tasks, the data is distributed according to a Bernoulli distribution.

2.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

Is the following statement True or False? Multi-class and Multi-label classification are the same thing.

True

False

Answer explanation

Multiclass classification - classification task where each instance needs to be assigned to one of two or more classes.

Mulitlabel classification - assign each instance a set of target labels. Think of this as predicting a series of properties of an instance which are not mutually exclusive.

3.

MULTIPLE CHOICE QUESTION

2 mins • 1 pt

The shape of the weight matrix, W of a Neural network linear layer is (x,y). A forward pass through this layer can be represented as follows:

Z=XW

Where X is the batched data of dimensions (batch size, input features) and W is the weight matrix. Select the correct assignments for (x, y)

x = batch size, y = number input features to that layer

x = number input features to that layer, y =batch size

x = input features, y = number of neurons in that layer

x = batch size, y = number of neurons in that layer

none of the above

Answer explanation

Media Image

Every neuron in a fully connected neural layer is connected to every input via a weight. This creates a vector of dimensions equal to the input layer dimensions for every neuron in the current layer. Resulting in a stack of these vectors as deep as the number of neurons in the current layer. For Z=XW the neighbouring dimensions need to match.

4.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

If a neural network has a single output neuron, then the model may be used for:

Binary classification

Regression

Binary classification or regression

None of these

5.

MULTIPLE CHOICE QUESTION

5 mins • 1 pt

Media Image

Here we have a computational graph representing a series of operations. The green text (text above the line) represents the forward pass (i.e. the values at each stage in the graph during forward propagation). The red values (the values which are positioned underneath the lines) represent gradient signals which have been passed back down the computational graph after some loss as been calculated. Calculate the missing gradients a, b and c using backpropagation (slides 20-31 onwards).

a= -0.2, b= 0.2, c=0.4

a= 0.2, b= -0.2, c=0.4

a= 0.4, b= 0.4, c=0.2

a= -0.2, b= -0.2, c=0.2

Answer explanation

Media Image

The key with this computational graph is to apply the chain rule at each node. Please see the attached solution for more information.

6.

MULTIPLE SELECT QUESTION

3 mins • 1 pt

Media Image

Which of the following are correct statements about the models which created the plots i) and ii)?

The model which created plot i) is demonstrating high bias and a higher capacity model is needed

The model which created plot ii) is demonstrating high bias and a higher capacity model is needed

The model which created plot ii) is demonstrating high variance and requires some form of regularisation to avoid over fitting.

The model which created plot i) is demonstrating high variance and requires some form of regularisation to avoid over fitting.

Answer explanation

i) is demonstrating high bias as 1.) the validation and training errors are high and at a similar level 2.) as we add more training data the training error increases. This happens as the capacity of the model limits its ability to fit all the modalities present in the larger dataset.

ii) is demonstrating high variance as the validation error is much higher than the training error. The model has overfit to spurious patterns in the training set which are not representative of the true distribution (represented by the validation set)