Комментарии:
Thanks Brandon
Very cool!
That was an amazing explanation. Very clear with the help of images and symbols.
ОтветитьSubtitles are auto-generated with lots of mistakes.
ОтветитьThanks a lot for this straight to the point explanation. Really learnt a lot.
ОтветитьThis is simply amazing and so clear.
ОтветитьWatching in 2x speed, even then it makes perfect sense! 🙏
ОтветитьAmazing Explanation!!
ОтветитьThe best Intro To RNNS and LSTMs I have seen!
Ответитьim a student major in data science from taiwan and i wanna say thank you, i got more to know about LSTM after watching ur video. Much appreciate sir!
ОтветитьWaffles for dinner? yuck
ОтветитьHi , i need some help here
why we decide to make the next hidden state = the long memory after filter it ? why not the next hidden layer not = the long memory (Ct)
subscribed! TYSSM
ОтветитьExceptionally good, the best I've seen in this subject.
ОтветитьI don't usually place any comment like this, but this is extraordinary :) So easy to understand :) Thank you :)
ОтветитьSuperb! Thanks!
ОтветитьThe best video on LSTM and RNN I've seen. Thank you so much!
ОтветитьAre the neural network units in an LSTM trained independently or simultaneously through backpropagation? I am assuming the latter and the functions of each unit are simply learned through their placement. For example, the "selection" NN unit is just named that from its placement and we don't do anything special during the training phase to make it a "selection" unit.
Ответитьyou lost me at former and latter
ОтветитьHey Thanks Bradon, Complexity put forth in Layman's language, just loved it!!!
Ответить