0 votes
in Deep Learning by

Why Sigmoid or Tanh is not preferred to be used as the activation function in the hidden layer of the neural network?

1 Answer

0 votes
by

A common problem with Tanh or Sigmoid functions is that they saturate. Once saturated, the learning algorithms cannot adapt to the weights and enhance the performance of the model. Thus, Sigmoid or Tanh activation functions prevent the neural network from learning effectively leading to a vanishing gradient problem. The vanishing gradient problem can be addressed with the use of Rectified Linear Activation Function (ReLu) instead of sigmoid and Tanh. Alt_text

Related questions

+1 vote
asked Jul 17, 2020 in Deep Learning by RShastri
0 votes
asked Aug 4, 2021 in Ingression Deep Learning by Robindeniel
...