MIT 6.S094: Recurrent Neural Networks for Steering Through Time

MIT 6.S094: Recurrent Neural Networks for Steering Through Time

Lex Fridman

7 лет назад

137,917 Просмотров

Ссылки и html тэги не поддерживаются


Комментарии:

@MM-vw1ck
@MM-vw1ck - 31.12.2023 22:07

My god, Lex is so lost on this lecture. It's almost like he forgot what he wanted to say when building the presentation.

Ответить
@justinchen207
@justinchen207 - 16.12.2023 09:52

god DAMN he was chunky. Rlly Came a long way

Ответить
@handokosupeno5425
@handokosupeno5425 - 15.11.2023 07:13

Amazing lecture

Ответить
@Phantastischphil
@Phantastischphil - 26.03.2023 16:59

Is there a playlist for the lectures leading up to this?

Ответить
@NIRVU
@NIRVU - 24.03.2023 08:32

How U can talk about something in complex detail yet sound like an illiterate r*tard is amazing Mr. Friedman

Ответить
@nemis123
@nemis123 - 16.03.2023 21:03

This gave me so much understanding. Thank you for uploading!

Ответить
@misssurreal2602
@misssurreal2602 - 25.02.2023 09:14

So you are familiar with vanilla... of course...

Ответить
@ajiteshsingh3764
@ajiteshsingh3764 - 08.09.2022 02:11

Beautiful, Just Beautiful.

Ответить
@quintrankid8045
@quintrankid8045 - 31.07.2022 06:26

Parameter tuning can't be taught? But it can be learned? I wonder if that would be a useful thing to apply ML to?

Ответить
@norik1616
@norik1616 - 09.07.2022 15:15

You look very cute compared to 2022 here 🤩

Ответить
@pratik245
@pratik245 - 23.12.2021 04:12

Can someone tell me why those vanishing gradients or exploding gradients occur, since i am such a dumb guy, i want to correlate it to nature.

Ответить
@pratik245
@pratik245 - 23.12.2021 04:04

Where is that sky view 360?

Ответить
@carlossegura403
@carlossegura403 - 13.09.2021 23:04

I still prefer the LSTM (for accuracy) | GRU (for speed) over the Transformer's architecture for both; their ability to learn long-dependencies and their simplicity.

Ответить
@tayg8370
@tayg8370 - 22.04.2021 06:36

Bruh I have no idea what he’s talking about but I’m some how interested

Ответить
@usf5914
@usf5914 - 21.04.2021 17:02

this presentation was bad. 1000 thumbs up was ...?

Ответить
@anthonybyne2724
@anthonybyne2724 - 27.10.2020 12:05

I've never listened to anyone before without understanding anything at all. It's fascinating for me watching with zero understanding. I'm literally just listening to his words... 😂

Ответить
@funduk89
@funduk89 - 27.10.2020 00:43

I think before introducing the backprop it is a good idea to start with the forward mode

Ответить
@TheAcujlGamer
@TheAcujlGamer - 06.10.2020 19:13

Aweosome!

Ответить
@stefanfaiciuc4560
@stefanfaiciuc4560 - 27.09.2020 17:56

Thanks so much for this, Lex. Your lecture was how I finally understood how RNNs work and it helped me to successfully complete my university thesis back in 2017. It's funny how I came across you again through Joe Rogan and your podcast and figured it's the same dude that helped me through college. Hope you get to be the one that builds robots better than anybody else in the world.

Ответить
@Anon_life
@Anon_life - 25.09.2020 13:44

I love this! <3

Ответить
@subramaniannk3364
@subramaniannk3364 - 16.07.2020 10:50

Is it that most explanations given for RNN are top-down and most explanations for CNN are bottom-up?

Ответить
@ravenmoore3399
@ravenmoore3399 - 10.07.2020 05:43

Vanilla is also what we call squares people who prefer missionary position are vanilla lol just saying

Ответить
@chrysr7900
@chrysr7900 - 15.05.2020 12:47

Ooh I recognize what's on the blackboard! It's the spherical coordinate system...

Ответить
@FezanRafique
@FezanRafique - 20.11.2019 05:55

Brilliant

Ответить
@user-ri3xr8vf1n
@user-ri3xr8vf1n - 21.10.2019 18:28

amazing lecture

Ответить
@physicsguy877
@physicsguy877 - 19.08.2019 06:08

It is extremely concerning that these students are not expected to know calculus cold. There is no such thing as, "but I understand the concepts". You use basic technical skill to check your understanding of concepts, so without knowing your abcs, you will tend to convince yourself of things that aren't true. There is a lot of democratizing technology out there now where you don't need to know what's going on "under-the-hood", but without at least some knowledge, all you will be able to do is press buttons and make graphs.

Ответить
@snehalbhartiya6724
@snehalbhartiya6724 - 30.07.2019 13:43

"If you enjoyed the video" he said ! Maybe you should rethink what you say.

Ответить
@yasar723
@yasar723 - 16.07.2019 17:17

Perfect at 1.25 speed

Ответить
@aidenstill7179
@aidenstill7179 - 04.06.2019 08:43

Please answer me. What do I need to know to create my own Python deep learning framework? What are the books and courses to get knowledge for this?

Ответить
@Gannicus99
@Gannicus99 - 14.02.2019 03:54

Some, hopefully helpful for the audience, remarks:
1. You need a lot of data. Depends. A lot of unlabeled data helps — to model the world. Then u need very little supervised data. Easy problems require little data. Hard or badly defined tasks require a lot of data. You can always pick an easier to solve proxy objective and use data augmentation.
2. rnns are dynamic length. Hard Set sequence lengths are for speed since:
sentences come at different lengths. So u cant create batches, unless you set a hard sequence length and then train same length sentences together in a batch, or fill up sentences that are too short by padding.
If you batch sentence you can compute on them in parallel.

Now of you are trying to predict relations between consecutive sentences, batching/ parallelization would not update the weights after each sentence, but on all of them at once — making it near impossible to learn inter (between) sentence relations but allowing the net to learn intra (within) sentence relations.

Tip: read karparthys blog on rnns not the Colah one. Karpathys is more detailed allowing you to really grasp what an rnn does. An lstm is „just“ and rnn with attention/gating.

Hope this helps, even if some concepts are very high level.

Ответить
@Constantinesis
@Constantinesis - 15.12.2018 19:54

Hello, thanks for uploading these lectures! Can LSTM networks integrate symbolic constructs in natural language learning? Can it help computers understand the relationship between language structure and real world? For example if I ask "Why is only raining outside? " It should know that the roof stops the rain falling inside. I have a feeling that we are mostly teaching the algorithm to interact with us, in some kind of smart language simulation but at it's core it doesn't really understand the meaning and relationships between words. Do you know some online references towards this?

Ответить
@vman049
@vman049 - 11.09.2017 20:15

Great video, but I wish there was more math and a more thorough explanation of BPTT and the vanishing gradient problem.

Ответить
@AhmedThabit99
@AhmedThabit99 - 08.09.2017 04:07

why their is no subtitle for this vedio ?!

Ответить
@mikashen5053
@mikashen5053 - 13.08.2017 16:37

Hey,Lex. Really great video! But as English is not my mother tongue, sometimes it's difficult to understand the video very well, it would be nice if you can turn on the cc subtitle options, thanks!

Ответить
@flamingxombie
@flamingxombie - 06.08.2017 20:29

Good video!

Ответить
@dongwang8848
@dongwang8848 - 02.08.2017 06:26

Great course with huge content in it, I am curious whether the Guest Talks are available too.

Ответить
@yangyang5447
@yangyang5447 - 24.06.2017 07:02

Thanks for the course. I learned a lot from it. Thanks!

Ответить
@JamesNortonSound
@JamesNortonSound - 05.06.2017 00:26

Am I the only person who thought that the video compression makes his shadow look like a low resolution shadow map...? Awesome content, great for getting into ML!

A quick question regarding LSTM's, why do we need a separate way of saying 'this information isn't important, I don't want to update my weights'. Doesn't gradient descent already take care of this? That is, if we happen to see a feature that is unimportant, won't we compute low gradients, thus telling us we don't need to move the weight by much? Why doesn't that work here?

Ответить
@feixyzliu5432
@feixyzliu5432 - 30.04.2017 06:27

Really cool course! Hi Lex, why only this the fourth lecture has no subtitle (Other 4 lectures do have)? Could you please upload one? Thank you.

Ответить
@SamirKhan-os2pr
@SamirKhan-os2pr - 27.04.2017 07:40

Academic by day... bouncer by night

Ответить
@abdulelahalkhoraif4495
@abdulelahalkhoraif4495 - 18.04.2017 13:27

I have to say you are very talented for teaching very complex topics. Thank you so much MIT for choosing such a brilliant presenter.

Ответить
@Graywolf116
@Graywolf116 - 09.04.2017 05:03

Hi Lex, are the Traffic/Tesla competitions still running? I see they're up on the site but with no end-dates. Were the prizes only for ver. 1.0 or also for the ver 1.1 currently up?

Ответить