Everything new and interesting in Machine Learning, Deep Learning, Data Science, & Artificial Intelligence. Hoping to build a community of data science geeks and talk about future tech! Projects demos and more! Subscribe for awesome videos :)
Ajay you are starting a cult man!!! May God bless you.
@atamir83393 күн бұрын
loved your explanation bro, earned yourself a sub
@scott79483 күн бұрын
Nice vid, but are you going to show us how to use a transformer to forecast into the future across multiple timesteps. I'm hanging out for that video. Cheers
@RakeshKambojVinayak4 күн бұрын
Thnx
@hassanahmedkhan38345 күн бұрын
Excellent Explanation, hats off.
@abhiksarkar92786 күн бұрын
This is most appropriate and concise explanation of attention
@matankri6 күн бұрын
Great content and well explained, but the excessive graphic visualizations like the moving characters is extremely distracting. Less is better.
@prodbreeze7 күн бұрын
B A B
@mahrym9847 күн бұрын
Great summary thank you🙏🏾
@alirezasalehabadi14227 күн бұрын
Thank you.
@prodbreeze7 күн бұрын
YOU HAVE MADE ME ACTALLY LIKE ML DL for the first time
@sammykmfmaths74688 күн бұрын
Please the video margin Is truncated 😢😢
@Sickkkkiddddd8 күн бұрын
How is information preserved with BN? If there was some pattern in the data pre-normalisation, wouldn't that be lost after batch norm?
@maga98528 күн бұрын
Hi @CodeEmpoium, Thank you very much for the video. It is very useful and well done. I just have one clarification. When you draw the Gaussian distribution corresponding to the point Xi (01:10-01:30), it seems that you make the peak of the curve coincide with the intersection of the interpolating line. Usually, it doesn’t happen this way; the peak is at the intersection of the vertical line passing through Xi and the interpolating line. As shown in this image: towardsdatascience.com/probabilistic-interpretation-of-linear-regression-clearly-explained-d3b9ba26823b Is that correct?
@maxsch.23678 күн бұрын
absolute banger! well done
@telugumoviesfunnycuts53109 күн бұрын
Could not get anything from this. too complex
@verendrakharkwal36429 күн бұрын
How is it 20:1 for alexnet 👀
@jayktharwani982210 күн бұрын
great explanation.
@adelAKAdude10 күн бұрын
great video thanks question ... in the third question ... how do sample subset of keys, queries "depending on importance"
@lucianofloripa12310 күн бұрын
Good explanation!
@vyax11 күн бұрын
lmk if they say
@surajrao972912 күн бұрын
the best video so far
@sillystuff624712 күн бұрын
you're a clear, calm explainer.
@darshh.poetry219312 күн бұрын
Nice explanation
@myhofficiel461213 күн бұрын
well explained , you made it look really easy !
@algorithmo13413 күн бұрын
what book is that
@vishwasprabhu404715 күн бұрын
Vsauce plugin 😂
@poorinspirit832215 күн бұрын
1-B 2-ABCD 3-C are these answers correct?
@poorinspirit832215 күн бұрын
1-CD 2-D 3-D
@poorinspirit832215 күн бұрын
1 - B, 2 - B, 3 - C. Are these answers correct?
@BruceOnder16 күн бұрын
That's not Pong.
@poorinspirit832216 күн бұрын
1 - A 2 - ABC 3 - A Are these answers correct?
@thangarajr-qw6wy17 күн бұрын
(1) supervised fine-tuning (SFT), (2) reward model (RM) training, and (3) reinforcement learning via proximal policy optimization (PPO) on this reward model explain me
@DevanshSagar-cy8kp17 күн бұрын
Great work ❤
@thvk9817 күн бұрын
I have seen so many transformers videos but this one is outstanding, I also want to request you to make a video on vision transformers too❤
@marcdelabarreraibardalet475417 күн бұрын
Nice video, well explained. Question, why would I use one or the other? Are there advantages or disadvantages?
@user-jy5pu6bg5p17 күн бұрын
Is it your voice or what ai you use ?
@AymaneArfaoui18 күн бұрын
what does x and y represent in the graph you use to show the cats and dog points ?
@Shrimant-ub4ul18 күн бұрын
thank u so much
@ParthivShah19 күн бұрын
Really Appreciate Your Efforts. Love from Gujarat India.
@swethanandyala19 күн бұрын
The best explanations on transformers that i have seen!
@einsteinwallah219 күн бұрын
according to dr. c. k. raju calculus was stolen from india
@samson670719 күн бұрын
You are my new favorite channel
@aswinselva0319 күн бұрын
The video is informative and good. but stop saying quiz time in an annoying way
@benjaminlin838620 күн бұрын
Super clear, thanks!
@cphrase21 күн бұрын
1997
@VideoDarslar21 күн бұрын
Easy
@Piku.aka.Priyanka21 күн бұрын
1. either of MSE or MAE as both are used as loss functions for regression. 2.A,B,C 3.A are these answers correct?
@Eysh200923 күн бұрын
Valeu!
@CodeEmporium22 күн бұрын
Thanks so much for the donation! Glad you liked this content!
Пікірлер
Thank you for this. What are U, V, and W at 8:44?
Ajay you are starting a cult man!!! May God bless you.
loved your explanation bro, earned yourself a sub
Nice vid, but are you going to show us how to use a transformer to forecast into the future across multiple timesteps. I'm hanging out for that video. Cheers
Thnx
Excellent Explanation, hats off.
This is most appropriate and concise explanation of attention
Great content and well explained, but the excessive graphic visualizations like the moving characters is extremely distracting. Less is better.
B A B
Great summary thank you🙏🏾
Thank you.
YOU HAVE MADE ME ACTALLY LIKE ML DL for the first time
Please the video margin Is truncated 😢😢
How is information preserved with BN? If there was some pattern in the data pre-normalisation, wouldn't that be lost after batch norm?
Hi @CodeEmpoium, Thank you very much for the video. It is very useful and well done. I just have one clarification. When you draw the Gaussian distribution corresponding to the point Xi (01:10-01:30), it seems that you make the peak of the curve coincide with the intersection of the interpolating line. Usually, it doesn’t happen this way; the peak is at the intersection of the vertical line passing through Xi and the interpolating line. As shown in this image: towardsdatascience.com/probabilistic-interpretation-of-linear-regression-clearly-explained-d3b9ba26823b Is that correct?
absolute banger! well done
Could not get anything from this. too complex
How is it 20:1 for alexnet 👀
great explanation.
great video thanks question ... in the third question ... how do sample subset of keys, queries "depending on importance"
Good explanation!
lmk if they say
the best video so far
you're a clear, calm explainer.
Nice explanation
well explained , you made it look really easy !
what book is that
Vsauce plugin 😂
1-B 2-ABCD 3-C are these answers correct?
1-CD 2-D 3-D
1 - B, 2 - B, 3 - C. Are these answers correct?
That's not Pong.
1 - A 2 - ABC 3 - A Are these answers correct?
(1) supervised fine-tuning (SFT), (2) reward model (RM) training, and (3) reinforcement learning via proximal policy optimization (PPO) on this reward model explain me
Great work ❤
I have seen so many transformers videos but this one is outstanding, I also want to request you to make a video on vision transformers too❤
Nice video, well explained. Question, why would I use one or the other? Are there advantages or disadvantages?
Is it your voice or what ai you use ?
what does x and y represent in the graph you use to show the cats and dog points ?
thank u so much
Really Appreciate Your Efforts. Love from Gujarat India.
The best explanations on transformers that i have seen!
according to dr. c. k. raju calculus was stolen from india
You are my new favorite channel
The video is informative and good. but stop saying quiz time in an annoying way
Super clear, thanks!
1997
Easy
1. either of MSE or MAE as both are used as loss functions for regression. 2.A,B,C 3.A are these answers correct?
Valeu!
Thanks so much for the donation! Glad you liked this content!