← Back to Quizzes

Neural Networks and Initialization

In the context of Univariate Logistic Regression with one input layer and one output layer, which activation function σ(z)\sigma(z) is defined as 11+exp(z)\frac{1}{1 + \exp(-z)}?
According to the lecture slides, what distinguishes a "Shallow" Neural Network from a standard "Machine Learning" (Logistic Regression) model?
If a neural network has an input layer a[0]a^{[0]}, a hidden layer a[1]a^{[1]}, and an output layer a[2]a^{[2]}, which layer represents the final prediction?
In a Multivariate Logistic Regression model, if the output activation a[1]0.5a^{[1]} \geq 0.5, what is the predicted class y^\hat{y}?
Consider a Shallow Neural Network with 3 input neurons (nx=3n_x=3) and 2 hidden neurons (nh=2n_h=2). According to the matrix form presented in the slides, what are the dimensions of the weight matrix W[1]W^{[1]}?
Following the previous question (3 inputs, 2 hidden neurons), if the output layer has 1 neuron (no=1n_o=1), what are the dimensions of the weight matrix W[2]W^{[2]}?
Why is it recommended to initialize weights randomly rather than setting them all to zero?
Which Python/Numpy command is suggested in the slides to initialize weights with small random numbers from a Gaussian distribution?
What is the primary purpose of setting a random seed (e.g., torch.manual_seed(seed)) in neural network training?
In the context of the slides, when is a hidden layer particularly useful?