Videos

Activation Functions in a Neural Network explained



deeplizard

In this video, we explain the concept of activation functions in a neural network and show how to specify activation functions in code with Keras.

Check out the corresponding blog and other resources for this video at: http://deeplizard.com/learn/video/m0pIlLfpXWE

Follow deeplizard on Twitter:
https://twitter.com/deeplizard

Follow deeplizard on Steemit:
https://steemit.com/@deeplizard

Become a patron:
https://www.patreon.com/deeplizard

Support deeplizard:
Bitcoin: 1AFgm3fLTiG5pNPgnfkKdsktgxLCMYpxCN
Litecoin: LTZ2AUGpDmFm85y89PFFvVR5QmfX6Rfzg3
Ether: 0x9105cd0ecbc921ad19f6d5f9dd249735da8269ef

Recommended books:
The Most Human Human: What Artificial Intelligence Teaches Us About Being Alive: http://amzn.to/2GtjKqu

Source

Similar Posts

29 thoughts on “Activation Functions in a Neural Network explained
  1. In relu, More positive it is, more activated the node is. For, 1 it'll be activated, for 3 it'll be activated also..! Than what's the difference between different positive values? Node is gonna activated anyway

  2. 1:11 where does the six come from? is it a trainable value like a bias, or is it the number of nodes in the active layer, or is it always six? In other words, why is the sigmoid's input limited to six?

  3. Hi, I would like to know which of the independent variables is much significant or has the highest impact on the dependent variable. My model is a 9-5-1 MLP which I have extracted its weights and biases. However, my concern now is how to use those weights to determine the most relevant input to the least relevant so I can rank. Thank you.

  4. Hi, I found this from somewhere "Because of the horizontal line in ReLu for when x<0, the gradient can go towards 0 . For activations in that region of ReLu, gradient will be 0 causing the weights to not get adjusted during descent. Hence, those neurons die and not respond making part of the NN passive"

    I don't understand why a gradient of 0, (gradient in this context seems to refer to the gradient of the relu function rather than the gradient of the loss function) will not allow for weights to get adjusted. Please help!

  5. This is totally wrong way to explain why we need activation function. We hide behind the biological explanation of neurons to explain why activation function is needed.

  6. Activation function happens on the second half of a node. So it might be miss leading to show all the weight pointing to the next hidden layer.

  7. without a doubt you are one of the best instructor I have ever seen. I absolutely love your channel. well done! and thank you so much for your great help. You saved my life

Comments are closed.

WP2Social Auto Publish Powered By : XYZScripts.com