In neural networks a activation function is a function that
defines a threshold that makes a node of a neural network to
activate. One example of a such activation function is the
sigmoid function
$$
\sigma(x) = \dfrac{1}{1 + e^{-x}}.
$$
(Aggarwal, 2023; Rojas, 1996).
When training a neural network, activation function’s derivative is
needed, but what it is for the sigmoid function?
it can be obvious that applying the quotient rule of
derivatives (E.g. see Grossman, 1986), i.e. for a given two
functions $f : \mathbb{R} \rightarrow \mathbb{R}$ and $g :
\mathbb{R} \rightarrow \mathbb{R}$
Aggarwal, C. C. (2023). Neural Networks and Deep Learning - A Textbook. 2nd ed.
Grossman, S. I. (1986). Calculus of One Variable. 2nd ed.
Rojas, R. (1996). Neural Networks - A Systematic Introduction. Springer.