Convolutional Neural Networks complete course is currently being offered by deeplearning.ai through Coursera platform and is Course 4 of 5 in the Deep Learning Specialization.
About this Course
In the fourth course of the Deep Learning Specialization, you will understand how computer vision has evolved and become familiar with its exciting applications such as autonomous driving, face recognition, reading radiology images, and more.
By the end, you will be able to build a convolutional neural network, including recent variations such as residual networks; apply convolutional networks to visual detection and recognition tasks; and use neural style transfer to generate art and apply these algorithms to a variety of image, video, and other 2D or 3D data.
The Deep Learning Specialization is our foundational program that will help you understand the capabilities, challenges, and consequences of deep learning and prepare you to participate in the development of leading-edge AI technology. It provides a pathway for you to gain the knowledge and skills to apply machine learning to your work, level up your technical career, and take the definitive step in the world of AI.
Also Check: How to Apply for Coursera Financial Aid

Convolutional Neural Networks Week 1 Quiz Answers
Q1. What do you think applying this filter to a grayscale
image will do?
- Detect
45 degree edges
- Detect
horizontal edges
- Detect
image contrast
- Detect
vertical edges
Q2. Suppose your input is a 300 by 300 color (RGB) image,
and you are not using a convolutional network. If the first hidden layer has
100 neurons, each one fully connected to the input, how many parameters does
this hidden layer have (including the bias parameters)?
- 9,000,100
- 9,000,001
- 27,000,001
- 27,000,100
Q3. Suppose your input is a 300 by 300 color (RGB) image,
and you use a convolutional layer with 100 filters that are each 5×5. How many
parameters does this hidden layer have (including the bias parameters)?
- 7500
- 2600
- 7600
- 2501
Q4. You have an input volume that is 63x63x16, and convolve
it with 32 filters that are each 7×7, using a stride of 2 and no padding. What
is the output volume?
- 29x29x16
- 16x16x32
- 16x16x16
- 29x29x32
Q5. You have an input volume that is 15x15x8, and pad it
using “pad=2.” What is the dimension of the resulting volume (after padding)?
- 19x19x12
- 19x19x8
- 17x17x10
- 17x17x8
Q6. You have an input volume that is 63x63x16, and convolve
it with 32 filters that are each 7×7, and stride of 1. You want to use a “same”
convolution. What is the padding?
- 3
- 7
- 2
- 1
Q7. You have an input volume that is 32x32x16, and apply max
pooling with a stride of 2 and a filter size of 2. What is the output volume?
- 16x16x16
- 32x32x8
- 15x15x16
- 16x16x8
Q8. Because pooling layers do not have parameters, they do
not affect the backpropagation (derivatives) calculation.
- True
- False
Q9. In lecture we talked about “parameter sharing” as a
benefit of using convolutional networks. Which of the following statements
about parameter sharing in ConvNets are true? (Check all that apply.)
- It
reduces the total number of parameters, thus reducing overfitting.
- It
allows a feature detector to be used in multiple locations throughout the
whole input image/input volume.
- It
allows parameters learned for one task to be shared even for a different
task (transfer learning).
- It
allows gradient descent to set many of the parameters to zero, thus making
the connections sparse.
Q10. In lecture we talked about “sparsity of connections” as
a benefit of using convolutional layers. What does this mean?
- Each
layer in a convolutional network is connected only to two other layers
- Regularization
causes gradient descent to set many of the parameters to zero.
- Each
activation in the next layer depends on only a small number of activations
from the previous layer.
- Each
filter is connected to every channel in the previous layer.
Convolutional Neural Networks Week 2 Quiz Answers
- nH and nW decrease, while nC increases
- nH and nW increases, while nC decreases
- nH and nW increases, while nC also increases
- nH and nW decreases, while nC also decreases
- FC layers in the last few layers
- FC layers in the first few layers
- Multiple CONV layers followed by a POOL layer
- Multiple POOL layers followed by a CONV layer
- True
- False
- True
- False
- 0 and a[l], respectively
- a[l] and 0, respectively
- z[l] and a[l], respectively
- 0 and z[l+1], respectively
- The skip-connections compute a complex non-linear function of the input to pass to a deeper layer in the network.
- The skip-connection makes it easy for the network to learn an identity mapping between the input and the output within the ResNet block.
- A ResNet with L layers would have on the order of L2 skip connections in total.
- Using a skip-connection helps the gradient to backpropagate and thus helps you to train deeper networks
- 2
- 1
- 17
- 4097
- You can use a pooling layer to reduce nH, nW, and nC.
- You can use a pooling layer to reduce nH, nW, but not nC.
- You can use a 1x1 convolutional layer to reduce nH, nW, and nC.
- You can use a 1x1 convolutional layer to reduce nC but not nH, nW.
- Inception networks incorporates a variety of network architectures (similar to dropout, which randomly chooses a network architecture on each step) and thus has a similar regularizing effect as dropout.
- Inception blocks usually use 1x1 convolutions to reduce the input data volume’s size before applying 3x3 and 5x5 convolutions.
- Making an inception network deeper (by stacking more inception blocks together) should not hurt training set performance.
- A single inception block allows the network to use a combination of 1x1, 3x3, 5x5 convolutions and pooling.
- It is a convenient way to get working an implementation of a complex ConvNet architecture.
- Parameters trained for one computer vision task are often useful as pretraining for other computer vision tasks.
- The same techniques for winning computer vision competitions, such as using multiple crops at test time, are widely used in practical deployments (or production system deployments) of ConvNets.
- A model trained for one computer vision task can usually be used to perform data augmentation even for a different computer vision task.
- y = [1, 0.3, 0.7, 0.5, 0.5, 1, 0, 0]
- y = [0, 0.2, 0.4, 0.5, 0.5, 0, 1, 0]
- y = [1, 0.7, 0.5, 0.3, 0.3, 0, 1, 0]
- y = [1, 0.3, 0.7, 0.3, 0.3, 0, 1, 0]
- y = [1, 0.3, 0.7, 0.5, 0.5, 0, 1, 0]
- y = [0, ?, ?, ?, ?, 0, 0, 0]
- y = [0, ?, ?, ?, ?, ?, ?, ?]
- y = [1, ?, ?, ?, ?, ?, ?, ?]
- y = [?, ?, ?, ?, ?, ?, ?, ?]
- y = [1, ?, ?, ?, ?, 0, 0, 0]
- Logistic unit, b_x and b_y
- Logistic unit, b_x, b_y, b_h, b_w
- Logistic unit, b_x, b_y, b_h(since b_w = b_h)
- Logistic unit (for classifying if there is a soft-drink can in the image)
- N
- N2
- 2N
- 3N
- True
- False
- True
- False
- True
- False
- 8
- 1/6
- 1/9
- 1/10
- None of the above
- 3
- 4
- 5
- 6
- 7
- 19x19x(5x20)
- 19x19x(5x25)
- 19x19x(20x25)
- 19x19x(25x20)
- True
- False
- We need to solve a one-shot learning problem.
- This allows us to learn to recognize a new person given just a single image of that person.
- Given how few images we have per person, we need to apply transfer learning.
- This allows us to learn to predict a person’s identity using a softmax output unit, where the number of classes equals the number of persons in the database plus 1 (for the final “not in database” class).
- True
- False
- max(∣∣f(A)−f(N)∣∣2 − ∣∣f(A)−f(P)∣∣2−α,0)
- max(||f(A)-f(P)||2 - ||f(A)-f(N)||2 + α, 0)
- max(||f(A)-f(N)||2 - ||f(A)-f(P)||2 + α, 0)
- max(||f(A)-f(P)||2 - ||f(A)-f(N)||2 - α, 0)
- True
- False
- True
- False
- True
- False
- True
- False
- The neural network parameters
- The regularization parameters
- The pixel values of the content image CC
- The pixel values of the generated image GG
- 30x30x30x16
- 30x30x30x32
- Undefined: This convolution step is impossible and cannot be performed because the dimensions specified don’t match up.
Post a Comment