Tutorial 7- Vanishing Gradient Problem

Vanishing Gradient Problem occurs when we try to train a Neural Network model using Gradient based optimization techniques. Vanishing Gradient Problem was actually a major problem 10 years back to train a Deep neural Network Model due to the long training process and the degraded accuracy of the Model.
Below are the various playlist created on ML,Data Science and Deep Learning. Please subscribe and support the channel. Happy Learning!
Deep Learning Playlist: • Tutorial 1- Introducti...
Data Science Projects playlist: • Generative Adversarial...
NLP playlist: • Natural Language Proce...
Statistics Playlist: • Population vs Sample i...
Feature Engineering playlist: • Feature Engineering in...
Computer Vision playlist: • OpenCV Installation | ...
Data Science Interview Question playlist: • Complete Life Cycle of...
You can buy my book on Finance with Machine Learning and Deep Learning from the below url
amazon url: www.amazon.in/Hands-Python-Fi...
🙏🙏🙏🙏🙏🙏🙏🙏
YOU JUST NEED TO DO
3 THINGS to support my channel
LIKE
SHARE
&
SUBSCRIBE
TO MY KZread CHANNEL

Пікірлер: 196

  • @kumarpiyush2169
    @kumarpiyush21694 жыл бұрын

    HI Krish.. dL/dW'11= should be [dL/dO21. dO21/dO11. dO11/dW'11] + [dL/dO21. dO21/dO12. dO12/dW'11] as per the last chain rule illustration. Please confirm

  • @rahuldey6369

    @rahuldey6369

    3 жыл бұрын

    ...but O12 is independent of W11,in that case won't the 2nd term be zero?

  • @RETHICKPAVANSE

    @RETHICKPAVANSE

    3 жыл бұрын

    wrong bruh

  • @ayushprakash3890

    @ayushprakash3890

    3 жыл бұрын

    we don't have the second term

  • @Ajamitjain

    @Ajamitjain

    3 жыл бұрын

    Can anyone clarify this? I too have this question.

  • @grahamfernando8775

    @grahamfernando8775

    3 жыл бұрын

    @@Ajamitjain dL/dW'11= should be [dL/dO21. dO21/dO11. dO11/dW'11]

  • @Vinay1272
    @Vinay1272 Жыл бұрын

    I have been taking a well-known world-class course on AI and ML since the past 2 years and none of the lecturers have made me so interested in any topic as much as you have in this video. This is probably the first time I have sat through a 15-minute lecture without distracting myself. What I realise now is that I didn't lack motivation or interest, nor that I was lazy - I just did not have lecturers whose teaching inspired me enough to take interest in the topics, yours did. You have explained the vanishing gradient problem so very well and clear. It shows how strong your concepts are and how knowledgeable you are. Thank you for putting out your content here and sharing your knowledge with us. I am so glad I found your channel. Subscribed forever.

  • @Xnaarkhoo
    @Xnaarkhoo3 жыл бұрын

    many years ago in the college I was enjoy watching videos from IIT - before the mooc area, India had and still have many good teachers ! It brings me joy to see that again. Seems Indians have a gene of pedagogy

  • @tosint
    @tosint4 жыл бұрын

    I hardly comment on videos, but this is a gem. One of the best videos explaining vanishing gradients problems.

  • @mahabir05
    @mahabir054 жыл бұрын

    I like how you explain and end your class "never give up " It very encouraging

  • @manishsharma2211

    @manishsharma2211

    3 жыл бұрын

    Yes

  • @aidenaslam5639
    @aidenaslam56394 жыл бұрын

    Great stuff! Finally understand this. Also loved it when you dropped the board eraser

  • @PeyiOyelo
    @PeyiOyelo4 жыл бұрын

    Sir or As my Indian Friends say, "Sar", you are a very good teacher and thank you for explaining this topic. It makes a lot of sense. I can also see that you're very passionate however, the passion kind of makes you speed up the explanation a bit making it a bit hard to understand sometimes. I am also very guilty of this when I try to explain things that I love. Regardless, thank you very much for this and the playlist. I'm subscribed ✅

  • @amc8437

    @amc8437

    3 жыл бұрын

    Consider reducing playback speed.

  • @ltoco4415
    @ltoco44154 жыл бұрын

    Thank you sir for making this misleading concept crystal clear. Your knowledge is GOD level 🙌

  • @marijatosic217
    @marijatosic2173 жыл бұрын

    I am amazed by the level of energy you have! Thank you :)

  • @benvelloor
    @benvelloor4 жыл бұрын

    Very well explained. I can't thank you enough for clearing all my doubts!

  • @rushikeshmore8890
    @rushikeshmore88904 жыл бұрын

    Kudos sir ,am working as data analyst read lots of blogs , watched videos but today i cleared the concept . Thanks for The all stuff

  • @manujakothiyal3745
    @manujakothiyal37454 жыл бұрын

    Thank you so much. The amount of effort you put is commendable.

  • @sapnilpatel1645
    @sapnilpatel1645 Жыл бұрын

    so far best explanation about vanishing gradient.

  • @gultengorhan2306
    @gultengorhan23062 жыл бұрын

    You are teaching better than many other people in this field.

  • @elielberra2867
    @elielberra2867 Жыл бұрын

    Thank you for all the effort you put into your explanations, they are very clear!

  • @vikrantchouhan9908
    @vikrantchouhan99082 жыл бұрын

    Kudos to your genuine efforts. One needs sincere efforts to ensure that the viewers are able to understand things clearly and those efforts are visible in your videos. Kudos!!! :)

  • @al3bda
    @al3bda3 жыл бұрын

    oh my god you are a good teacher i really fall in love how you explain and simplify things

  • @sumeetseth22
    @sumeetseth224 жыл бұрын

    Love your videos, I have watched and taken many courses but no one is as good as you

  • @koraymelihyatagan8111
    @koraymelihyatagan81112 жыл бұрын

    Thank you very much, I was wandering around the internet to find such an explanatory video.

  • @lekjov6170
    @lekjov61704 жыл бұрын

    I just want to add this mathematically, the derivative of the sigmoid function can be defined as: *derSigmoid = x * (1-x)* As Krish Naik well said, we have our maximum when *x=0.5*, giving us back: *derSigmoid = 0.5 * (1-0.5) --------> derSigmoid = 0.25* That's the reason the derivative of the sigmoid function can't be higher than 0.25

  • @ektamarwaha5941

    @ektamarwaha5941

    4 жыл бұрын

    COOL

  • @thepsych3

    @thepsych3

    4 жыл бұрын

    cool

  • @tvfamily6210

    @tvfamily6210

    4 жыл бұрын

    should be: derSigmoid(x) = Sigmoid(x)[1-Sigmoid(x)], and we know it reaches maximum at x=0. Plugging in: Sigmoid(0)=1/(1+e^(-0))=1/2=0.5, thus derSigmoid(0)=0.5*[1-0.5]=0.25

  • @benvelloor

    @benvelloor

    4 жыл бұрын

    @@tvfamily6210 Thank you!

  • @est9949

    @est9949

    3 жыл бұрын

    I'm still confused. The weight w should be in here somewhere. This seems to be missing w.

  • @deepthic6336
    @deepthic63364 жыл бұрын

    I must say this, normally I am kinda person who prefers to study on own and crack it. Never used to listen to any of the lectures till date because I just don't understand and I dislike the way they explain without passion(not all though). But, you are a gem and I can see the passion in your lectures. You are the best Krish Naik. I appreciate it and thank you.

  • @MauiRivera
    @MauiRivera3 жыл бұрын

    I like the way you explain things, making them easy to understand.

  • @classictremonti7997
    @classictremonti79973 жыл бұрын

    Krish...you rock brother!! Keep up the amazing work!

  • @piyalikarmakar5979
    @piyalikarmakar59792 жыл бұрын

    One of the best vedio on clarifying Vanishing Gradient problem..Thank you sir..

  • @venkatshan4050
    @venkatshan40502 жыл бұрын

    Marana mass explanation🔥🔥. Simple and very clearly said.

  • @meanuj1
    @meanuj15 жыл бұрын

    Nice presentation..so much helpful...

  • @bhavikdudhrejiya4478
    @bhavikdudhrejiya44784 жыл бұрын

    Very nice way to explain. Learned from this video- 1. Getting the error (Actual Output - Model Output)^2 2. Now We have to reduce an error i.e Backpropagation, We have to find a new weight or a new variable 3. Finding New Weight = Old weight x Changes in the weight 4. Change in the Weight = Learning rate x d(error / old weight) 5. After getting a new weight is as equals to old weight due to derivate of Sigmoid ranging between 0 to 0.25 so there is no update in a new weight 6. This is a vanishing gradient

  • @MrSmarthunky
    @MrSmarthunky4 жыл бұрын

    Krish.. You are earning a lot of Good Karmas by posting such excellent videos. Good work!

  • @swapwill
    @swapwill4 жыл бұрын

    The way you explain is just awesome

  • @mittalparikh6252
    @mittalparikh62523 жыл бұрын

    Overall got the idea, that you are trying to convey. Great work

  • @daniele5540
    @daniele55404 жыл бұрын

    Great tutorial man! Thank you!

  • @himanshubhusanrath2492
    @himanshubhusanrath24922 жыл бұрын

    One of the best explanations of vanishing gradient problem. Thank you so much @KrishNaik

  • @benoitmialet9842
    @benoitmialet98422 жыл бұрын

    Thank you so much, great quality content.

  • @classictremonti7997
    @classictremonti79973 жыл бұрын

    So happy I found this channel! I would have cried if I found it and it was given in Hindi (or any other language than English)!!!!!

  • @shmoqe
    @shmoqe2 жыл бұрын

    Great explanation, Thank you!

  • @vishaljhaveri6176
    @vishaljhaveri61762 жыл бұрын

    Thank you, Krish SIr. Nice explanation.

  • @b0nnibell_
    @b0nnibell_4 жыл бұрын

    you sir made neural network so much fun!

  • @sunnysavita9071
    @sunnysavita90714 жыл бұрын

    your videos are very helpful ,good job and good work keep it up...

  • @nabeelhasan6593
    @nabeelhasan65932 жыл бұрын

    Very nice video sir , you explained very well the inner intricacies of this problem

  • @skiran5129
    @skiran51292 жыл бұрын

    I'm lucky to see this wonderful class.. Tq..

  • @yoyomemory6825
    @yoyomemory68253 жыл бұрын

    Very clear explanation, thanks for the upload.. :)

  • @MsRAJDIP
    @MsRAJDIP5 жыл бұрын

    Tommorow I have interview, clearing all my doubts from all your videos 😊

  • @YashSharma-es3lr
    @YashSharma-es3lr3 жыл бұрын

    very simple and nice explanation . I understand it in first time only

  • @satyadeepbehera2841
    @satyadeepbehera28414 жыл бұрын

    Appreciate your way of teaching which answers fundamental questions.. This "derivative of sigmoid ranging from 0 to 0.25" concept was nowhere mentioned.. thanks for clearing the basics...

  • @mittalparikh6252

    @mittalparikh6252

    3 жыл бұрын

    Look for Mathematics for Deep Learning. It will help

  • @naresh8198
    @naresh8198 Жыл бұрын

    crystal clear explanation !

  • @nazgulzholmagambetova1198
    @nazgulzholmagambetova1198 Жыл бұрын

    great video! thank you so much!

  • @nola8028
    @nola80282 жыл бұрын

    You just earned a +1 subscriber ^_^ Thank you very much for the clear and educative video

  • @faribataghinezhad
    @faribataghinezhad2 жыл бұрын

    Thank you sir for your amazing video. that was great for me.

  • @hiteshyerekar9810
    @hiteshyerekar98105 жыл бұрын

    Nice video Krish.Please make practicle based video on gradient decent,CNN,RNN.

  • @neelanshuchoudhary536
    @neelanshuchoudhary5364 жыл бұрын

    very nice explanation,,great :)

  • @nirmalroy1738
    @nirmalroy17384 жыл бұрын

    super video...extremely well explained.

  • @melikad2768
    @melikad27683 жыл бұрын

    Thank youuuu, its really great:)

  • @BalaguruGupta
    @BalaguruGupta3 жыл бұрын

    Thanks a lot sir for the wonderful explanation :)

  • @narsingh2801
    @narsingh28014 жыл бұрын

    You are just amazing. Thnx

  • @ambreenfatimah194
    @ambreenfatimah1943 жыл бұрын

    Helped a lot....thanks

  • @krishj8011
    @krishj80113 жыл бұрын

    Very nice series... 👍

  • @dhananjayrawat317
    @dhananjayrawat3174 жыл бұрын

    best explanation. Thanks man

  • @tonnysaha7676
    @tonnysaha76763 жыл бұрын

    Thank you thank you thank you sir infinite times🙏.

  • @arunmeghani1667
    @arunmeghani16673 жыл бұрын

    great video and great explanation

  • @adityashewale7983
    @adityashewale7983 Жыл бұрын

    hats off to you sir,Your explanation is top level, THnak you so much for guiding us...

  • @DEVRAJ-np2og

    @DEVRAJ-np2og

    15 күн бұрын

    do u completed his full playlist?

  • @yousufborno3875
    @yousufborno38754 жыл бұрын

    You should get Oscar for your teaching skills.

  • @GunjanGrunge
    @GunjanGrunge2 жыл бұрын

    that was very well explained

  • @sunnysavita9071
    @sunnysavita90714 жыл бұрын

    very good explanation.

  • @abhinavkaushik6817
    @abhinavkaushik68172 жыл бұрын

    Thank you so much for this

  • @abdulqadar9580
    @abdulqadar95802 жыл бұрын

    Great efforts Sir

  • @khiderbillal9961
    @khiderbillal99613 жыл бұрын

    thanks sir you really hepled me

  • @maheshsonawane8737
    @maheshsonawane8737 Жыл бұрын

    Very nice now i understand why weights doesn't update in RNN. The main point is derivative of sigmoid is between 0 and 0.25. Vanishing gradient is associated with only sigmoid function. 👋👋👋👋👋👋👋👋👋👋👋👋

  • @nikunjlahoti9704
    @nikunjlahoti97042 жыл бұрын

    Great Lecture

  • @narayanjha3488
    @narayanjha34885 жыл бұрын

    This video is amazing and you are amazing teacher thanks for sharing such amazing information Btw where are you from banglore?

  • @gaurawbhalekar2006
    @gaurawbhalekar20064 жыл бұрын

    excellent explanation sir

  • @prerakchoksi2379
    @prerakchoksi23794 жыл бұрын

    I am doing deep learning specialization, feeling that this is much better than that

  • @AA-yk8zi
    @AA-yk8zi3 жыл бұрын

    Thank you so much

  • @salimtheone
    @salimtheone Жыл бұрын

    very well explained 100/100

  • @feeham
    @feeham2 жыл бұрын

    Thank you !!

  • @skviknesh
    @skviknesh3 жыл бұрын

    I understood it. Thanks for the great tutorial! My query is: weight vanishes when respect to more layers. When new weight ~= old weight result becomes useless. what would the O/P of that model look like (or) will we even achieve global minima??

  • @ilyoskhujayorov8498
    @ilyoskhujayorov84983 жыл бұрын

    Thank you !

  • @Thriver21
    @Thriver21 Жыл бұрын

    nice explanation.

  • @AnirbanDasgupta
    @AnirbanDasgupta3 жыл бұрын

    excellent video

  • @ArthurCor-ts2bg
    @ArthurCor-ts2bg4 жыл бұрын

    Excellent 👌

  • @hokapokas
    @hokapokas5 жыл бұрын

    Good job bro as usual... Keep up the good work.. I had a request of making a video on implementing back propagation. Please make a video for it.

  • @krishnaik06

    @krishnaik06

    5 жыл бұрын

    Already the video has been made.please have a look on my deep learning playlist

  • @hokapokas

    @hokapokas

    5 жыл бұрын

    @@krishnaik06 I have seen that video but it's not implemented in python.. If you have a notebook you can refer me to pls

  • @krishnaik06

    @krishnaik06

    5 жыл бұрын

    With respect to implementation with python please wait till I upload some more videos

  • @naimamushfika1167
    @naimamushfika1167 Жыл бұрын

    nice explanation

  • @amitdebnath2207
    @amitdebnath22072 ай бұрын

    Hats Off Brother

  • @Haraharavlogs
    @Haraharavlogs6 ай бұрын

    you are legend nayak sir

  • @manikosuru5712
    @manikosuru57125 жыл бұрын

    As usual extremely good outstanding... And a small request can expect this DP in coding(python) in future??

  • @krishnaik06

    @krishnaik06

    5 жыл бұрын

    Yes definitely

  • @aaryankangte6734
    @aaryankangte67342 жыл бұрын

    Sir thank u for teaching us all the concepts from basics but just one request is that if there is a mistake in ur videos then pls rectify it as it confuses a lot of people who watch these videos as not everyone sees the comment section and they just blindly belive what u say. Therefore pls look into this.

  • @ganeshkharad
    @ganeshkharad4 жыл бұрын

    nice explaination

  • @anusuiyatiwari1800
    @anusuiyatiwari18003 жыл бұрын

    Very interesting

  • @susmitvengurlekar
    @susmitvengurlekar3 жыл бұрын

    Understood completely! If weights hardly change, no point in training and training. But I have got a question, where can I use this knowledge and understanding I just acquired ?

  • @lalithavanik5022
    @lalithavanik50223 жыл бұрын

    Nice expalnation sir

  • @muhammadarslankahloon7519
    @muhammadarslankahloon75193 жыл бұрын

    Hello sir, why the chain rule explained in this video is different from the very last chain rule video. kindly clearly me and thanks for such an amazing series on deep learning.

  • @naughtyrana4591
    @naughtyrana45914 жыл бұрын

    Guruvar ko pranam🙏

  • @rayyankhattak544
    @rayyankhattak544 Жыл бұрын

    Great-

  • @raj4624
    @raj46242 жыл бұрын

    superb

  • @sandipansarkar9211
    @sandipansarkar92114 жыл бұрын

    Thanks krish .Video was superb but I am having apprehension I might get lost somewhere .Please provide some reading reference regrading this topic considering as a beginner.Cheers

  • @shahidabbas9448
    @shahidabbas94484 жыл бұрын

    Sir i'm really confusing about the actual y value please can you tell about that. i thought it would be our input value but here input value is so many with one predicted output

  • @suryagunda4038
    @suryagunda40383 жыл бұрын

    May god bless you ..

  • @varayush
    @varayush3 жыл бұрын

    @krish: thanks for the wonderful lessons on the neural network. may I request you to correct the equation using some text box on the video as this will have intact information that you would like to pass on

  • @magicalflute
    @magicalflute4 жыл бұрын

    Very well explained. Vanishing gradient problem as per my understanding is that, it is not able to perform the optimizer job (to reduce the loss) as old weight and new weights will be almost equal. Please correct me, if i am wrong. Thanks!!

  • @sekharpink
    @sekharpink5 жыл бұрын

    Derivative of loss with respect to w11 dash you specified incorrectly, u missed derivative of loss with respect to o21 in the equation. Please correct me if iam wrong.

  • @sekharpink

    @sekharpink

    5 жыл бұрын

    Please reply

  • @ramleo1461

    @ramleo1461

    4 жыл бұрын

    Evn I hv this doubt

  • @krishnaik06

    @krishnaik06

    4 жыл бұрын

    Apologies for the delay...I just checked the video and yes I have missed that part.

  • @ramleo1461

    @ramleo1461

    4 жыл бұрын

    @@krishnaik06Hey!, U dnt hv to apologise, on the contrary u r dng us a favour by uploading these useful videos, I was a bit confused and wanted to clear my doubt that all, thank you for the videos... Keep up the good work!!

  • @rajatchakraborty2058

    @rajatchakraborty2058

    4 жыл бұрын

    @@krishnaik06 I think you have also missed the w12 part in the derivative. Please correct me if I am wrong

  • @aishwaryaharidas2100
    @aishwaryaharidas21004 жыл бұрын

    Should we again add bias to the product of the output from the hidden layer O11, O12 and weights W4, W5?