What is weight decay?

  1. What is weight decay?
    •  A technique to avoid vanishing gradient by imposing a ceiling on the values of the weights.
    •  A regularization technique (such as L2 regularization) that results in gradient descent shrinking the weights on every iteration.
    •  The process of gradually decreasing the learning rate during training.
    •  Gradual corruption of the weights in the neural network if it is trained on noisy data.

Get All Week Quiz Answer:

Improving Deep Neural Networks: Hyperparameter Tuning, Regularization and Optimization Coursera Quiz Answer

Similar Posts