0 votes
in Deep Learning by
When it comes to training an artificial neural network, what could be the reason why the loss doesn't decrease in a few epochs?

1 Answer

0 votes
by

Some of the reasons why the loss doesn't decrease after a few Epochs are:

a) The model is under-fitting the training data.

b) The learning rate of the model is large.

c) The initialization is not proper (like all the weights initialized with 0 doesn't make the network learn any function)

d) The Regularisation hyper-parameter is quite large.

e). The classic case of vanishing gradients

Related questions

+1 vote
asked Jul 19, 2023 in Deep Learning by SakshiSharma
0 votes
asked Jul 21, 2023 in Deep Learning by rahuljain1
...