Notice: This page requires JavaScript to function properly.
Please enable JavaScript in your browser settings or update your browser.
Aprenda Sigmoid and Tanh Activations | Section
Deep Learning for Sequential Data

bookSigmoid and Tanh Activations

Deslize para mostrar o menu

The sigmoid and tanh activation functions are explored, as they play a crucial role in the functioning of RNNs.

Note
Definition

The sigmoid and tanh functions transform inputs into outputs, enabling the model to make predictions.

Sigmoid
  • Sigmoid activation: the sigmoid function maps input values to an output range between 0 and 1. It is commonly used in binary classification tasks, as its output can be interpreted as a probability. However, it suffers from the vanishing gradient problem when the input values are very large or very small;
  • Tanh activation: the tanh function is similar to the sigmoid but maps the input values to an output range between -1 and 1. It helps center the data around zero, which can aid learning. Despite its benefits, it also suffers from the vanishing gradient problem in certain situations;
  • Working of sigmoid and tanh: both functions work by squashing the input values into a bounded range. The primary difference lies in their output range: sigmoid (0 to 1) vs. tanh (-1 to 1), which affects how the network processes and updates the information.

In the next chapter, we will look at how these activation functions play a role in LSTM networks and how they help overcome some of the limitations of standard RNNs.

question mark

What is the output range of the sigmoid activation function?

Selecione a resposta correta

Tudo estava claro?

Como podemos melhorá-lo?

Obrigado pelo seu feedback!

Seção 1. Capítulo 7

Pergunte à IA

expand

Pergunte à IA

ChatGPT

Pergunte o que quiser ou experimente uma das perguntas sugeridas para iniciar nosso bate-papo

Seção 1. Capítulo 7
some-alt