0
0
Computer Visionml~20 mins

Why architecture design impacts performance in Computer Vision - Challenge Your Understanding

Choose your learning style9 modes available
Challenge - 5 Problems
🎖️
Architecture Mastery
Get all challenges correct to earn this badge!
Test your skills under time pressure!
🧠 Conceptual
intermediate
2:00remaining
How does increasing model depth affect performance?

In convolutional neural networks, what is the most common effect of increasing the number of layers (depth) on model performance?

AIt reduces performance because deeper models cannot learn complex features.
BIt always improves performance by capturing more complex features without any drawbacks.
CIt can improve performance by learning complex features but may cause vanishing gradients and overfitting.
DIt has no effect on performance since only the number of neurons matters.
Attempts:
2 left
💡 Hint

Think about what happens when a model becomes very deep and how training might be affected.

Predict Output
intermediate
2:00remaining
Output shape after convolution and pooling layers

Given the following PyTorch model snippet, what is the output shape after the last layer?

Computer Vision
import torch
import torch.nn as nn

class SimpleCNN(nn.Module):
    def __init__(self):
        super().__init__()
        self.conv1 = nn.Conv2d(3, 16, kernel_size=3, padding=1)
        self.pool = nn.MaxPool2d(2, 2)
        self.conv2 = nn.Conv2d(16, 32, kernel_size=3, padding=1)
    def forward(self, x):
        x = self.pool(torch.relu(self.conv1(x)))
        x = self.pool(torch.relu(self.conv2(x)))
        return x

model = SimpleCNN()
input_tensor = torch.randn(1, 3, 64, 64)
output = model(input_tensor)
output.shape
Atorch.Size([1, 32, 8, 8])
Btorch.Size([1, 16, 8, 8])
Ctorch.Size([1, 16, 16, 16])
Dtorch.Size([1, 32, 16, 16])
Attempts:
2 left
💡 Hint

Calculate the size after each convolution and pooling step.

Hyperparameter
advanced
2:00remaining
Choosing kernel size impact on feature extraction

How does increasing the convolution kernel size from 3x3 to 7x7 typically affect a CNN's ability to extract features?

ALarger kernels capture more global features but increase parameters and risk overfitting.
BLarger kernels always improve performance without any drawbacks.
CLarger kernels reduce the receptive field and limit feature extraction.
DKernel size does not affect feature extraction, only the number of filters matters.
Attempts:
2 left
💡 Hint

Think about how kernel size relates to the area of the image the filter sees.

Metrics
advanced
2:00remaining
Interpreting validation accuracy drop with deeper architecture

A CNN model with 10 layers achieves 85% validation accuracy. Increasing to 50 layers drops validation accuracy to 70%. What is the most likely reason?

AThe dataset is too large for the deeper model to learn.
BThe deeper model suffers from vanishing gradients and overfitting.
CThe deeper model is underfitting due to too few parameters.
DValidation accuracy always decreases with more layers.
Attempts:
2 left
💡 Hint

Consider training difficulties with very deep networks.

🔧 Debug
expert
3:00remaining
Identifying cause of exploding gradients in deep CNN

Given a deep CNN training with exploding gradients, which architectural choice is the most likely cause?

AUsing ReLU activation without batch normalization in a very deep network.
BUsing batch normalization and residual connections in the network.
CUsing small kernel sizes like 3x3 in all convolution layers.
DUsing dropout layers after every convolution.
Attempts:
2 left
💡 Hint

Think about what helps stabilize training in deep networks.