Add to Chrome
✅ The verified answer to this question is available below. Our community-reviewed solutions help you understand the material better.
Why might this network suffer from vanishing gradients?
Because softmax increases gradients
Because of softmax
Because 784 input features are too few
Because sigmoid activation causes small gradients in deep layers
Get Unlimited Answers To Exam Questions - Install Crowdly Extension Now!