MIT 6.S191 (2021): Recurrent Neural Networks

Поділитися
Вставка
  • Опубліковано 7 січ 2025

КОМЕНТАРІ • 159

  • @nilotpalmrinal5120
    @nilotpalmrinal5120 3 роки тому +157

    “The definition of genius is taking the complex and making it simple.” - A congruence of this quote and this lecture series defines the quality of Instructors. Thanks a ton to Alex and Ava!!! Thank you very much!

    • @nickstaresinic9933
      @nickstaresinic9933 3 роки тому +6

      Yes. You nailed it.

    • @BoomBaaamBoom
      @BoomBaaamBoom 3 роки тому +2

      Yes. It is nice to have such a high quality lecture available to the public. I am also impressed by the instructors. Go Ava!

  • @VikasKM
    @VikasKM 3 роки тому +12

    One of the clearest explanations on RNNs, LSTMs.

  • @lakshmisrinivas369
    @lakshmisrinivas369 3 роки тому +4

    I really liked the way LSTM concept is explained. The attention mechanism has been briefly described yet well explained. Thank you so much.

  • @normalhuman6260
    @normalhuman6260 2 роки тому

    Will be honest, this is probably the best lecture I have ever seen on DL. Most other lectures are so inaccessible and jargon filled that they fail to drive home the fundamentals. Kudos to Ava and Alexander

  • @cedricmanouan2333
    @cedricmanouan2333 3 роки тому +15

    Even if I’ve already watched previous lectures, I am watching these ones as it is the first time. Masterpiece 😭❤️🐐

  • @harunismail3044
    @harunismail3044 3 роки тому

    MIT gives a title of introduction to deep learning, but some people realize that it is quite deep rationale behind what they have sought for a long time ago. Thank you MIT. A great lecture.

  • @aaronwilliams1143
    @aaronwilliams1143 3 роки тому +5

    Incredible lecture, I had to pause halfway through just to absorb as much information as I could. Please keep these coming, I have a great aptitude for neural networks! This course is right up my alley :)

  • @otakudnp3880
    @otakudnp3880 3 роки тому +15

    Going from Attack on Titans to Deep Learning. What a week :)

  • @programmer1379
    @programmer1379 3 роки тому +14

    Clear intro. to RNNs building up intuition from the basic principles Loved the lecture !

  • @reandov
    @reandov 3 роки тому +11

    Thanks for the amazing class once again! Recurrent Neural Networks are very strong and important nowadays in our society, and the improvement and studies about them make a huge impact on this!

  • @prashantkumar7390
    @prashantkumar7390 2 роки тому +1

    Our instructor's flow is super smooth, no cap

  • @jonothan
    @jonothan 3 роки тому +7

    This is the best explanation of LSTMs I've seen!

  • @sree_haran
    @sree_haran 3 роки тому +6

    Thanks a lot for making these mit lectures public... I'm so happy to learn these.. it's all because of you 🤗

  • @DawidOhia
    @DawidOhia 3 роки тому +11

    Ava Soleimany has a really high level skills of knowledge explaining. Thanks for making these lectures public.

  • @isaacguerreiro3869
    @isaacguerreiro3869 3 роки тому +2

    I would like to thanks Alex and Ava. Have this content with this quality is priceless for someone that is trying to learn ML and DL by himself. Thank you for share this incredible class online for free.

  • @zigzag4273
    @zigzag4273 3 роки тому +1

    Ava really cleared the confusing bits of the internal workings of standard RNNs and LSTMs. Thanks.
    Thanks Ava and Alex.

  • @Jacob011
    @Jacob011 3 роки тому +6

    This lecture is perfection! I say that as a pedantic PhD 🙂. I can tell that a crap ton of work went into it.

  • @FernandoWittmann
    @FernandoWittmann 3 роки тому

    Just a minor correction in 20:12 which I couldn't find the reason anywhere: tf.keras.layers.SimpleRNN actually implements an output-to-output recurrence which is slightly different from the scratch model provided. This can be demonstrated by checking their weights and manually multiplying them. The output-to-output recurrence (when omitting the bias) would be self.h = self.W_xh*x --> output = tf.math.tanh(self.h+ previous_output*w_oo). If anyone knows why keras implements output-to-output instead of hidden-to-hidden, please let me know, I usually see in literature hidden-to-hidden.

    • @FernandoWittmann
      @FernandoWittmann 3 роки тому

      I find it to be more intuitive since it can be simplified to: output = tanh(W_xh*x+ output*w_oo), hence w_oo is a weight that is "memorizing" or "giving importance" to the previous outputs.

  • @ajaytaneja111
    @ajaytaneja111 3 роки тому +3

    I just wait all day through at office to get back home and watch this amazing series of lectures. Thank you Team @Alexander Amini

  • @mehdismaeili3743
    @mehdismaeili3743 2 роки тому

    سلام خانوم کتر.عالی بود.شما باعث افتخار ما ایرانیها همچنین مایه ی فخر MIT هستید.بی نظیر هستین.

  • @RamanVenu
    @RamanVenu 3 роки тому

    Kudos to Alex and Ava for taking the time to making these concepts so simple to consume! I'm definitely going to watch all the lectures for this course.

  • @prashanthvaidya
    @prashanthvaidya 3 роки тому +22

    Grateful to you profs and MIT! 💯
    What a wonderful introduction to the intuition behind RNNs. :)

  • @khalilrekik4542
    @khalilrekik4542 3 роки тому +31

    Going from WandaVision to Deep learning. What a weekend :D

  • @SamuelGuebo
    @SamuelGuebo 3 роки тому +1

    Many thanks for taking the time to produce and release this invaluable content. Cheers from Abidjan!

  • @rolandheinze7182
    @rolandheinze7182 3 роки тому +4

    Pretty sure I tried viewing this lecture series at least 2 years ago, and this format is much more understandable and digestible. Thank you MIT and lecturers/producers.

  • @andreasw665
    @andreasw665 3 роки тому +15

    Perfect timing for the weekend ;) can’t wait

  • @shriharimutalik3231
    @shriharimutalik3231 3 роки тому +2

    I watched all the lectures of 2020 in the lockdown , when I first I thought what I learnt in other course for 1 week is equivalent what I learn from a sngle lecture , its too cool to cover all these coccepts in an hour..

    • @bobpearsall7737
      @bobpearsall7737 3 роки тому

      at MIT, this is called learning by "drinking from a firehose". once you get used to it, other options seem tedious.

  • @rezamemarian4010
    @rezamemarian4010 Рік тому +1

    That was the best video I watched about RNN. Thank you 😊

  • @stephennfernandes
    @stephennfernandes 3 роки тому +2

    Perfect , loved the previous RNN lecture watched it over and over , couldn’t wait for this one

  • @quackdoescrck3009
    @quackdoescrck3009 3 роки тому +1

    i did two courses and know alot ... but my understanding about this becomes more and more clear... thanks for these videos

  • @ryan_chew97
    @ryan_chew97 3 роки тому +4

    it really shows - MIT = world class and nothing else

  • @drdr3496
    @drdr3496 Рік тому +1

    This lecture is clearer than the waters of the Caribbean. Fantastic.

  • @haenarashin3738
    @haenarashin3738 3 роки тому +3

    The best lecture, the best weekends!

  • @richarda1630
    @richarda1630 3 роки тому +2

    Thanks for your awesome explanation of LSTMs! can't wait to see the deep-dive into Transformers and how they compare! :)

  • @SenthilBalas
    @SenthilBalas 2 роки тому +5

    Love the 14:20 to 15:30 pseudo logic representation of RNN. Cant be represented more simpler !

  • @sidasrivas6004
    @sidasrivas6004 3 роки тому

    Thanks to these two Amazing Iranian young MIT professors!Ava and Alexander!

  • @annakozio9220
    @annakozio9220 3 роки тому +2

    Great lectures, the topics seem so simple to me now. Thank you Ava and Alexander!

  • @kaizhang9530
    @kaizhang9530 3 роки тому +4

    very well presented lecture, condensing much info into a one-hour session. Bravo!

  • @ngoziokereke3561
    @ngoziokereke3561 3 роки тому +1

    Once again, I really enjoyed this lecture. All the concepts are well explained, I am just about to start the lab session and feel quite excited to apply all I picked up from this lecture. Thank you!

  • @vimukthirandika872
    @vimukthirandika872 2 роки тому +1

    You guys are genius because things are simplified

  • @sabirhakro
    @sabirhakro 3 роки тому +3

    Very clearly explaining and Perfect teaching timing!

  • @jayasreemangalagiri971
    @jayasreemangalagiri971 3 роки тому +1

    "Beautiful" is the only word I can think of at the end of lecture #1

  • @anuragsangem1442
    @anuragsangem1442 3 роки тому +1

    Glad I came across this! perfect way to spend my weekend!
    Thanks.

  • @medhnhadush4320
    @medhnhadush4320 2 роки тому

    everything she said gets into my mind directly ,amazing lecture ! thank you very much

  • @sukantasengupta4025
    @sukantasengupta4025 3 роки тому +5

    Friday Evening.. Perfect 🕯️

  • @mojganrezakhani3403
    @mojganrezakhani3403 3 роки тому +1

    Thanks, Ava for your interesting presentation. I am proud of you as one Iranian girl.

  • @snivesz32
    @snivesz32 2 роки тому +1

    Question: So Ht is a single float being passed from one cell to the next? And what are the dimensions of the weight matrix W_hh? Because I assume it must learn something of value that is characterized by the relationship between the input at T-1 and T. So my question is, is the weight matrix W_hh dimensioned by the size of the vocabulary squared? If so what sort of intuition should I have about what H is encoding?

  • @AmineBenkirane-x1y
    @AmineBenkirane-x1y 4 місяці тому

    Well done on the lecturing skills improvement ! i could certainly feel that haha :) Thanks for the Lecture !

  • @DistortedV12
    @DistortedV12 3 роки тому +1

    Watching this right now, awesome Ava!

  • @fernandocutire4
    @fernandocutire4 3 роки тому +2

    Perfect timing !! Thanks for the upload

  • @_rd_kocaman
    @_rd_kocaman 3 роки тому +2

    This is the future of teaching and teaches you the future of technology

  • @Fun-io9kv
    @Fun-io9kv 3 роки тому +1

    Thanks a lot for the tutorials. Those were the best ones I 've ever seen about Deep Learning . Best Wishes

  • @macknightxu2199
    @macknightxu2199 3 роки тому

    16:50 why y only depends on h? doesn't y depend both on h and x? Does it mean that h(t) has considered the non-linear feature, and y(t) just considered as linear to h(t) is enough?

  • @jiholmes1251
    @jiholmes1251 3 роки тому +2

    amazing structure of the class, loved it

  • @haneulkim4902
    @haneulkim4902 3 роки тому +8

    Amazing lecture! Can't thank you enough. Thanks :)

  • @myspam4194
    @myspam4194 3 роки тому +2

    Thanks a lot for sharing, really appreciate the from scratch code part that's help me to test my understanding of what was described.

  • @Qaidi_804
    @Qaidi_804 2 роки тому +1

    Huge respect for such a great lecture. Thanks Ava

  • @emmanouilmorfiadakis118
    @emmanouilmorfiadakis118 2 роки тому

    Very good Representation and you can learn many things.Well done

  • @JP68
    @JP68 3 роки тому

    Amazing lecture!! Thanks for sharing! Congrats to professors and MIT!

  • @sumanthhabib8028
    @sumanthhabib8028 3 роки тому +1

    Amazing lectures, crystal clear explanation of concepts!

  • @jim8730
    @jim8730 3 роки тому

    The lecture was very well presented, and easy to understand. I learned a lot.
    Maybe because I'm still new to the subject, I found the explanation on Attention a little difficult to understand.
    Thank you very much!

  • @madhavanj92
    @madhavanj92 3 роки тому +3

    Is there anyone who is trying out the lab? @Alexander, consider setting up a forum where we could discuss issues /insights that we have while trying out the lab exercises

  • @k1llyah
    @k1llyah 3 роки тому +1

    Brilliant content, absolutely loving it. I am implementing sample networks using test datasets in R for the topics in the lectures. If anybody wants the files, let me know.
    The lecturers voice trembles a bit (not sure if nervous or exited), so some words of encouragement: you rock!

  • @luispitta2715
    @luispitta2715 3 роки тому

    This is a very complicated topic, but was a great lecture!! Congrats!!

  • @ganjarulez009
    @ganjarulez009 3 роки тому +14

    Imagine giving a full lecture without once saying "uhm", or something similar! Really high lecturing skills here

  • @uty6595
    @uty6595 3 роки тому +1

    Well done, best UA-cam channel I've subscribed. Please keep it up

  • @butterkaffee910
    @butterkaffee910 3 роки тому +1

    Attention is all we need

  • @EpsilonLessThan0
    @EpsilonLessThan0 2 роки тому +1

    Beautifully explained!

  • @usama57926
    @usama57926 3 роки тому +1

    Can't we use *ascii codes* for words rather than using *one hot encoding* . . . ❓ ❓ ❓

  • @mukherab
    @mukherab 3 роки тому +2

    Great detailed tutorials! thanks. Just a quick typo on Backpropagation through time slide (32 mins), shouldn't the Loss L_3 be L_t instead? i.e. corresponding to the t^th unit.

  • @gelarehnaseri8939
    @gelarehnaseri8939 3 роки тому +2

    Great content, presentation, and clarity! Thank you very much Ava.

  • @80amnesia
    @80amnesia 3 роки тому +1

    nice tutorial. I love the tutorials in this channel

  • @dhrumpelstiltskin
    @dhrumpelstiltskin 3 роки тому +3

    Is there an online community/forum for this course? I'm curious to see how others implemented solutions for the lab! (I implemented a solution for the lab but I don't think my implementation is very clean or efficient...)

  • @soyvoyager7148
    @soyvoyager7148 2 роки тому +1

    This lecture is statement of excellence🙏

  • @abhilekhadalal9126
    @abhilekhadalal9126 3 роки тому +2

    Hey, if we encounter problems while solving labs how do we approach it, is there any channel or forum or any medium to contact or solve our doubts??

  • @Abhi-qi6wm
    @Abhi-qi6wm 2 роки тому

    In the lecture she says that the backprop takes place through the cell state and the original pathway is left undisturbed, but if we don't backpropagate through the original pathway (i.e. xt & ht-1), how are the weights going to adjust to give a lower cost function value?

  • @mathewcraig
    @mathewcraig 3 роки тому +2

    Don’t have enough superlatives. Ava is amazing

  • @Rupadarshi-Ray
    @Rupadarshi-Ray 3 роки тому +1

    Past year's courses are further available for OCW than ever 👍🏻

  • @seyedalimirmohammadhoseiny9125
    @seyedalimirmohammadhoseiny9125 3 роки тому +4

    Thank you for this perfect lecture Alexander and Ava

  • @lucusekali5767
    @lucusekali5767 3 роки тому +2

    Thank you for our hero: UA-cam Algorithm

  • @fabiosouza1195
    @fabiosouza1195 3 роки тому +1

    Amazing Class! So enlightening about RNNs! Thank you for sharing all of these amazing classes!

  • @xinniu3145
    @xinniu3145 3 роки тому +2

    Thanks for providing this great course! Could anyone elaborate on the encoding bottleneck issue at 53:00? Is it only for LSTM or all RNNs?

  • @macknightxu2199
    @macknightxu2199 3 роки тому +1

    may I ask what's the differences between feed-forward and traditional neural networks? isn't traditional neural networks feed-forwarded?
    and RNN is not feed-forward neural networks, right? because it's recurrent neural network.

  • @HarpreetKaur-qq8rx
    @HarpreetKaur-qq8rx 3 роки тому

    Ava one question so are you saying that y at each time step is a word that is being predicted with each of the input word and hidden state.
    So what I mean by that at 0th time step word "I" is fed into the RNN and some word[basically a vector in the space is produced denoted by y0] is predicted and a hidden state is generated. Now this hidden state and the word "love" is fed into the 1st time step and some word[denoted by y1] is predicted and another hidden state is generated and this goes on until the last word is fed into the RNN. And here the time t is denoted by the number of words in the input sequence meaning 4 in this case since the sentence is composed of 4 words "I love recurrent neural"

  • @sinikishan1408
    @sinikishan1408 3 роки тому +1

    Great... awaiting the session

  • @aidamehdipour7072
    @aidamehdipour7072 3 роки тому +1

    Excellent presentation. Thanks.

  • @ravvbike
    @ravvbike 3 роки тому +4

    Hey guys, where can i have acess to the Lab exercises?

    • @AAmini
      @AAmini  3 роки тому +5

      github.com/aamini/introtodeeplearning

  • @Jootawallah
    @Jootawallah 3 роки тому

    Sorry, for gated RNNs, you talk about the self-state c_t. But how is c_0 initialised?

  • @adebowaledanieladebayo6957
    @adebowaledanieladebayo6957 3 роки тому +1

    My Friday night vibe. :D

  • @王吉-q4k
    @王吉-q4k 3 роки тому +1

    Thanks for uploading so great videos and I have one question about the list of the videos, are they in the right sequence or just placed randomly in the list?

  • @ninahristozova7198
    @ninahristozova7198 3 роки тому +1

    Truly amazing, thank you!

  • @iamshrutisneha
    @iamshrutisneha 3 роки тому

    Does RNN really able to handle variable length input sequences by its own or we need to provide the maximum_length parameter there so that the smaller inputs need to be padded??? Anyone please clarify this. Thanks

  • @madhavanjagannathan7292
    @madhavanjagannathan7292 3 роки тому +1

    Great stuff! Is there a forum where I can post clarifications on the lab?

  • @krishnachauhan2850
    @krishnachauhan2850 Рік тому

    Is the output equation is correct at 17.00 , is not it require activation there too?

  • @shashwatm67
    @shashwatm67 3 роки тому

    Eva mentioned "labs" a couple of times where RNN is implemented. Where can I get a video of the same?

  • @WARATEL_114_30
    @WARATEL_114_30 3 роки тому +3

    Can't wait

  • @hanners93
    @hanners93 2 роки тому

    Why is it not a network from hidden layer to output layer in slide 6.24 min in this video?

  • @mazharulhasan
    @mazharulhasan 3 роки тому +1

    Perfect lecture to understand...

  • @BLAXE-vm3ms
    @BLAXE-vm3ms 3 роки тому

    why do we use same weights for every time step in rnn why can't we use best posssible weight at that timestep?

  • @bouchelligamohamedhedi2747
    @bouchelligamohamedhedi2747 3 роки тому +1

    Thank you , that was awesome