deeplizard
In this video, we explain the concept of activation functions in a neural network and show how to specify activation functions in code with Keras.
Check out the corresponding blog and other resources for this video at: http://deeplizard.com/learn/video/m0pIlLfpXWE
Follow deeplizard on Twitter:
https://twitter.com/deeplizard
Follow deeplizard on Steemit:
https://steemit.com/@deeplizard
Become a patron:
https://www.patreon.com/deeplizard
Support deeplizard:
Bitcoin: 1AFgm3fLTiG5pNPgnfkKdsktgxLCMYpxCN
Litecoin: LTZ2AUGpDmFm85y89PFFvVR5QmfX6Rfzg3
Ether: 0x9105cd0ecbc921ad19f6d5f9dd249735da8269ef
Recommended books:
The Most Human Human: What Artificial Intelligence Teaches Us About Being Alive: http://amzn.to/2GtjKqu
Source
*Note, starting at 1:09, the denominator of the sigmoid function should be e^x+1 rather than e^(x+1).*
Machine Learning / Deep Learning Tutorials for Programmers playlist: https://www.youtube.com/playlist?list=PLZbbT5o_s2xq7LwI2y8_QtvuXZedL6tQU
Keras Machine Learning / Deep Learning Tutorial playlist: https://www.youtube.com/playlist?list=PLZbbT5o_s2xrwRnXk_yCPtnqqo4_u2YGL
Which programming language are you using?
Thanks a lot mam! This helped a lot
In relu, More positive it is, more activated the node is. For, 1 it'll be activated, for 3 it'll be activated also..! Than what's the difference between different positive values? Node is gonna activated anyway
Amazing tutorials!
Thank you so much
1:11 where does the six come from? is it a trainable value like a bias, or is it the number of nodes in the active layer, or is it always six? In other words, why is the sigmoid's input limited to six?
first and foremost,
thank you for amazing explain,
Hi, I would like to know which of the independent variables is much significant or has the highest impact on the dependent variable. My model is a 9-5-1 MLP which I have extracted its weights and biases. However, my concern now is how to use those weights to determine the most relevant input to the least relevant so I can rank. Thank you.
You're my favorite Machine Learning teacher, please keep making such videos.
You never explain WHY to use an activation function. You just showed us HOW it works.
great video Tnx!
Nice Explanation. What is softmax and why it is widely used in Deep learning instead of general activation functions?
You saved my tommorow's ML exam.
I've just started ML and thanks to you. Cuz of you I'm able to digest Coursera lectures practically. And finally you've earned a subscriber ?
Simple, but it helped me out a lot!
Thank you so much 🙂 Keep making such amazing videos please.
Does Deep learning with Keras need no skills?
One of my friend said that it's irrelevant to code with Keras :'-(
Hi, I found this from somewhere "Because of the horizontal line in ReLu for when x<0, the gradient can go towards 0 . For activations in that region of ReLu, gradient will be 0 causing the weights to not get adjusted during descent. Hence, those neurons die and not respond making part of the NN passive"
I don't understand why a gradient of 0, (gradient in this context seems to refer to the gradient of the relu function rather than the gradient of the loss function) will not allow for weights to get adjusted. Please help!
thank you very much
This is totally wrong way to explain why we need activation function. We hide behind the biological explanation of neurons to explain why activation function is needed.
The answer lies in linearity and non linearity.
Thank you for these videos! Its so easy to follow your explanations.
Activation function happens on the second half of a node. So it might be miss leading to show all the weight pointing to the next hidden layer.
without a doubt you are one of the best instructor I have ever seen. I absolutely love your channel. well done! and thank you so much for your great help. You saved my life
Perfect Explanation .
Extremely! Wonderful Explanation <3 it
Love your voice
can i access this jupyter notebook?