Challenge: Integrate Dropout and BatchNorm
Swipe to start coding
You will extend a simple neural network by integrating Dropout and Batch Normalization. Your goal is to correctly insert these layers into the architecture and perform a forward pass.
You are given:
- Input batch
x - A partially defined network class
- A forward method missing some components
Complete the following steps:
-
Add a Dropout layer after the first fully connected layer.
-
Add a BatchNorm layer immediately after Dropout.
-
Complete the forward pass so that the data flows through:
- Linear → ReLU → Dropout → BatchNorm → Linear
-
Ensure Dropout is used only during training (PyTorch handles this automatically).
After execution, the script prints the network output.
Lösning
Tack för dina kommentarer!
single
Fråga AI
Fråga AI
Fråga vad du vill eller prova någon av de föreslagna frågorna för att starta vårt samtal
Fantastiskt!
Completion betyg förbättrat till 8.33
Challenge: Integrate Dropout and BatchNorm
Svep för att visa menyn
Swipe to start coding
You will extend a simple neural network by integrating Dropout and Batch Normalization. Your goal is to correctly insert these layers into the architecture and perform a forward pass.
You are given:
- Input batch
x - A partially defined network class
- A forward method missing some components
Complete the following steps:
-
Add a Dropout layer after the first fully connected layer.
-
Add a BatchNorm layer immediately after Dropout.
-
Complete the forward pass so that the data flows through:
- Linear → ReLU → Dropout → BatchNorm → Linear
-
Ensure Dropout is used only during training (PyTorch handles this automatically).
After execution, the script prints the network output.
Lösning
Tack för dina kommentarer!
single