Suppose we have a 11-layer neural network which takes 10 hours to train on a GPU
with 32GB VRAM. At test time, it takes 5 seconds for single data point.
Now we change the architecture such that we add dropout after 2nd, 4th, 6th and
8th layer with rate 0.2 for each layer respectively.
What would be the testing time for this new architecture?
A Doubles
B Halves
C Same
D Can?t Say
Answers
Answered by
0
Answer:
A
Explanation:
bsnsndmdndnrmrbbfbbtvt
Answered by
0
Answer:
D
Explanation:
Because dropouts are random so they decrease training time
so it has to lower or the same time but definitely not halved
Similar questions