Softmax Regression (C2W3L08)

Take the Deep Learning Specialization: bit.ly/2xdG0Et
Check out all our courses: www.deeplearning.ai
Subscribe to The Batch, our weekly newsletter: www.deeplearning.ai/thebatch
Follow us:
Twitter: / deeplearningai_
Facebook: / deeplearninghq
Linkedin: / deeplearningai

Пікірлер: 48

  • @Jabrils
    @Jabrils6 жыл бұрын

    thank you master andrew, this was super to the point.

  • @roniquinonez9715

    @roniquinonez9715

    5 жыл бұрын

    Spotted a wild Jabril in his natural environment! *Attempts to say Hello*

  • @HZLTV
    @HZLTV Жыл бұрын

    I don't think I'll ever understand the maths behind this properly, but the fact that I even understood it *sort of* just proves how good he is at teaching... the visual example given helped a tonne

  • @mariusz2313
    @mariusz23136 жыл бұрын

    "If you can't explain it simply, you don't understand it well enough. " Albert Einstein You definitely know the topic perfectly well! Thanks!

  • @Jonn123save

    @Jonn123save

    6 жыл бұрын

    so true

  • @roboticsresources9680

    @roboticsresources9680

    6 жыл бұрын

    Actually , it was Richard Feynman who said it

  • @bornslippy9109

    @bornslippy9109

    6 жыл бұрын

    no einstein said it, feynman applied it perfectly

  • @est9949

    @est9949

    3 жыл бұрын

    This is a common logical fallacy: A implies B is not equivalent to B implies A.

  • @mortezaabdipour5584
    @mortezaabdipour55846 жыл бұрын

    Thank you, M.r Andrew, for always sharing your knowledge

  • @manuel783
    @manuel7833 жыл бұрын

    Clarifications about Softmax Regression Please note that at 4:30, the text for the softmax formulas mixes subscripts "j" and "i", when the subscript should just be the same (just "i") throughout the formula.

  • @jjjj_111
    @jjjj_1116 жыл бұрын

    Fantastic, I love how easy it was to understand the material that was presented. If you have a donation page, please let me know!

  • @Jack-dx7qb
    @Jack-dx7qb6 жыл бұрын

    So clear!

  • @mariabardas2568
    @mariabardas25684 жыл бұрын

    Great lesson !!!! Very useful.

  • @ziku8910
    @ziku89102 жыл бұрын

    That was very helpful, thank you!

  • @camilaonofri2624
    @camilaonofri26244 жыл бұрын

    If you use the sigmoid fx for a multiclass problem you have to make a decision boundary for each class against the others (1-vs-all algo), and then you get independent probabilities. How were the decision boundaries for the examples 10:33 calculated considering you had no hidden layers? (how were the line equations found?)

  • @84xyzabc
    @84xyzabc3 жыл бұрын

    I think, in the denominator, its t_j at 4:58

  • @stefandimeski8569
    @stefandimeski85696 жыл бұрын

    How were the decision boundaries for the examples 10:33 calculated? (how were the line equations found?)

  • @grez911

    @grez911

    6 жыл бұрын

    Make a grid, let's say 100 by 100 points, and in each point calculate the activation. It's actually how he do this in programming exercises.

  • @stefandimeski8569

    @stefandimeski8569

    6 жыл бұрын

    Thanks man! Then I suppose it's impossible to get a closed-form equation for the decision boundary? Can you please provide link to the video where he uses the method you described here?

  • @camilaonofri2624

    @camilaonofri2624

    4 жыл бұрын

    please let me know if you got it

  • @AdarshMahabubnagar
    @AdarshMahabubnagar2 жыл бұрын

    Is it possible to show the softmax activation function graphically? If so, please provide

  • @fjficm
    @fjficm2 жыл бұрын

    with "t" when you say normailsed, the probabilities would be different, wouldnt it? You would have to use 1 / sqr (t.t) in front of the 4x1 matrix and convert it into a unit matrix. Then use the dot products of the element to work out the probablities of each which will still work out to 1. Or is this wrong

  • @leagueofotters2774
    @leagueofotters27742 жыл бұрын

    Soft and soothing....kind of like the Bob Ross of machine learning.

  • @wolfisraging
    @wolfisraging6 жыл бұрын

    U r best

  • @wifi-YT
    @wifi-YT3 жыл бұрын

    At 9:32 forward, what’s the reason the decision boundaries are all linear, when the softmax function is NOT itself a linear function? The softmax function uses, after all, an e to the z function in its numerator, which function is certainly not linear! So, similarly, why is it that Andrew says at 11:16 that only when you add hidden layers do you end up with non-linear decision boundaries?

  • @yilmazbingol6953
    @yilmazbingol69535 жыл бұрын

    i think at time 4.30,sum indices should be j=0 to j=3. if i m wrong, pls correct me.

  • @mathias8137

    @mathias8137

    5 жыл бұрын

    This makes definitely more sense to me. The same applies for the sum written in 5:00

  • @scl144
    @scl1444 жыл бұрын

    번역해주신 분 누구신지 몰라서 동서남북으로 절했습니다. 너무 감사합니다.. 덕분에 훌륭한 강의 보고 갑니다

  • @benw4361
    @benw43614 жыл бұрын

    It seems like the largest number is still selected as the predicted solution i.e. 5, so I'm confused by what the purpose is of softmax when you could select the largest value instead? Wouldn't that effectively translate to the class with the largest probability anyway?

  • @michael3698bear

    @michael3698bear

    3 жыл бұрын

    Late reply but for anyone wonder. Yes you are correct, for prediction purposes it makes no difference (though still maybe you would like to see the "probabilities" generated), but you are correct that just the the max will be chosen as the "predicted solution". This however is not true for training. When training you need to be able to measure "how wrong" you were. This is where the softmax function comes in to give probabilities, which you can calculate loss from, and also calculate a derivative to update the weights.

  • @boratsagdiev6486
    @boratsagdiev64865 жыл бұрын

    It should be tj not ti at 4:33 right?

  • @mathiasgustum858

    @mathiasgustum858

    4 жыл бұрын

    yes :)

  • @lucylu2530

    @lucylu2530

    3 жыл бұрын

    I believe that under the sigma function it should be i=4

  • @cryhav0k2112
    @cryhav0k21126 жыл бұрын

    Where do the numbers in the Z vector come from?

  • @kunhongyu5053

    @kunhongyu5053

    6 жыл бұрын

    just assumption

  • @cryhav0k2112

    @cryhav0k2112

    6 жыл бұрын

    Kunhong YU sure, but i mean intuitively, what do they represent?

  • @kunhongyu5053

    @kunhongyu5053

    6 жыл бұрын

    Similar to simple Logistic Regression, Softmax just adds more output units rather than one. For logistic regression, output unit is just a 1 dimensional vector computing input X's linear "score", if it's larger than zero, then its label is 1 and vice versa. Softmax is like training multiple binary classifiers simultaneously, for a sample, each element in Z is also a "score", largest score denotes that sample may have corresponding label.

  • @DouglasDuhaime

    @DouglasDuhaime

    6 жыл бұрын

    @derrikbosse, the Z vector identified here has 3 arguments: W{L}, A{L-1}, and B{L}. W{L} is the vector of weights within the last layer of the network. A{L-1} is the vector of outputs from the penultimate layer of the network. B{L} is the bias vector from the last layer of the network. If none of this makes any sense, check out Professor Ng's earlier discussion of logistic regression, which is the simplest kind of neural network. That helped me make sense of this presentation: kzread.info/dash/bejne/mp6mu9Synqq9d6Q.html

  • @usf5914
    @usf59143 жыл бұрын

    4:03 (4, 1) or (1, 4)?

  • @gavin8535
    @gavin85353 жыл бұрын

    What is W^L at 3:56 ?

  • @aayushpaudel2379

    @aayushpaudel2379

    3 жыл бұрын

    Weight matrix for layer L , i.e the last layer.

  • @marcostavarez2702
    @marcostavarez27025 жыл бұрын

    could someone please explain the roles of the 'blocks-of-color' and the 'colored-in-circles' represent ?

  • @IrfanAhmad-od2sn

    @IrfanAhmad-od2sn

    5 жыл бұрын

    colored-in-circles are actual training data/values.After training model on training data,model predicted all decision boundaries.so color of blocks are blocks predicted by model.

  • @grez911
    @grez9116 жыл бұрын

    How do you calculate so quickly? I don't see a calculator on your table.

  • @chancychan7175

    @chancychan7175

    4 жыл бұрын

    Actually I cal it for MR.NG hh

  • @sandipansarkar9211
    @sandipansarkar92113 жыл бұрын

    Greqt exoplamantion.Ned to wtch again

  • @fatimahmath4819
    @fatimahmath48195 жыл бұрын

    ⚘⚘⚘⚘⚘