An introduction to Reinforcement Learning

Поділитися
Вставка
  • Опубліковано 20 гру 2024

КОМЕНТАРІ • 410

  • @DanielHernandez-rn6rp
    @DanielHernandez-rn6rp 6 років тому +291

    Love this guy. As an RL PhD student, your videos are golden.

    • @nikhillondhe5815
      @nikhillondhe5815 6 років тому +12

      RL PhD sounds so interesting!

    • @andres18m
      @andres18m 6 років тому +2

      Institute name?

    • @Ayanwesha
      @Ayanwesha 6 років тому +1

      hello..sir
      i am a grad stud
      can anyone tell me plzz if back propagation is necessary in supervised and unsupervised learning?or it is only used in reinforcement learning
      thanks

    • @hcgaron
      @hcgaron 5 років тому

      Ayanwesha 12345 yes, back propagation is used as a basis for gradient based methods of optimization

    • @ernie2111
      @ernie2111 5 років тому +3

      "RL PhD" didn't know such things exist lol

  • @atcer51
    @atcer51 Рік тому +1

    fiiiinnnaaaallly after tons of googling, I finally fund a USEFUL video that accually EXPLAINS how to reward the agent, and not just saying:
    'oh u just reward it'

  • @yuanyuansun3521
    @yuanyuansun3521 3 роки тому +32

    “If u only give it a positive reward when it successfully stacked a block, it’ll never get to see any of those reward” Only if my tutors realise this.

  • @denebvegaaltair1146
    @denebvegaaltair1146 2 роки тому +15

    Your videos have just the right amount of technical terms such that student engineers can learn something, and also the right amount of summary and rewording such that beginners can get a vague idea of concepts. Thank you so much

  • @rednassie1101
    @rednassie1101 4 роки тому +229

    People: ANN ARE TAKING OVER THE WORLD AND STUFF WILL NEVER BE THE SAME
    my horribly trained network on a cat: "dog"

    • @I_Lemaire
      @I_Lemaire 4 роки тому +1

      Could they help with the necessary government takeovers associated with COVID-19? Temporary command economies could be more efficient.

    • @revimfadli4666
      @revimfadli4666 4 роки тому +6

      UA-cam's bots: "Robot fighting is animal cruelty"

  • @SukhwinderSingh-fb9qw
    @SukhwinderSingh-fb9qw 6 років тому +66

    This was one of the best videos on RL that I have seen. Extremely informative. The way you explain things is awesome. Keep up the great work! Cheers man!

  • @davidfield5295
    @davidfield5295 6 років тому

    The misuse of 'literally' notwithstanding, this was an excellent video. Very clear and concise explanation.

  • @snippletrap
    @snippletrap 6 років тому +17

    The perils of reward shaping are well understood in a public policy context, where incentives can lead to "unintended consequences".

  • @funpy772
    @funpy772 3 роки тому +1

    Just wanted to tell you people.. this video is still awesome.

  • @cemgocer8185
    @cemgocer8185 4 роки тому +3

    Quality of the video is off the charts. Topics u have chosen to explain the field, the way u explain them and especially pointing the common misconceptions that make it harder for us to get into what AI really is... I'm sad that there is no superlike button. Rare to see videos of this quality and honesty

  • @Hyuts
    @Hyuts 5 років тому +17

    Explains in an elegant manner more than I have learned in half a semester of my AI college course.

  • @TheBeansChopper
    @TheBeansChopper 3 роки тому

    I think the comment section speaks for itself. This is a fantastic grasp of the basic concepts and issues with this technologies in such short time, without diving unnecessarily into formalism. Thanks :)

  • @floriandebrauwer9140
    @floriandebrauwer9140 5 років тому +2

    Thanks for your work ! I like the way you present such a complex field in a clear manner for poeple without any background. Thanks to you I know where to start in my learning journey !

  • @allamasadi7970
    @allamasadi7970 6 років тому +151

    Your channel deserves more views 👍

  • @rutexgreat3619
    @rutexgreat3619 2 місяці тому

    Very clear material, very clear representation, thank you for your time and video.

  • @shirishbajpai9486
    @shirishbajpai9486 Рік тому

    3:12 - Why reinforcement learning
    4:00 - RL framework
    4:30 - Policy Gradients
    5:37 - Training Policy network
    7:50 - Problem with policy gradient(credit assignment problem)
    9:25 - Where sparse reward setting fails
    11:00 - Reward shaping

  • @biiigates7381
    @biiigates7381 4 роки тому +1

    I've been learning AI for almost a year now and on all the channels I've spent with this is the best one. Very underrated! (btw its the first time i discovered this channel and I instantly subscribed)

    • @mundeepcool
      @mundeepcool 4 роки тому

      Same here, loved this video and I instantly subscribed... and also oh yeah yeah

  • @TY-un4no
    @TY-un4no 4 роки тому +1

    Complex stuff made simple and easy, this is a very good intro video to RL. Starting to learn RL for work and your video gave me a great starting point, thank you!

  • @Lilowillow42
    @Lilowillow42 3 роки тому +1

    Just wanted you to know that in my university course for introduction to AI our professor recommended your videos for machine learning. Your explanation is highly enjoyable and informative. Thank you!

  • @aanex2005
    @aanex2005 5 років тому

    I have no idea about RL but your video has given me a good jump start. Thanks man

  • @PriyanshuGupta-hf2hm
    @PriyanshuGupta-hf2hm 3 роки тому

    You explained so well that I understood each and everything in your video. I am overjoyed!

  • @punitpalial
    @punitpalial 3 роки тому

    Here the notes from the video, LEARNZY (please ignore the timestamps, they are not accurate)
    01:57 : Peter Abeel gave a demonstration of robots doing all the mundane tasks of the house like cleaning, cooking, and bringing a bottle of beer. It showed our remarkable achievements in the field of robotics
    We are sufficiently advanced(mechanically and in hardware essence) to build a robot capable of doing complex actions but the reason we aren't able to make terminator-like robots is that we still haven't embedded intelligence into these robots. So creating intelligent robots is a software problem, not a hardware problem
    02:03 : Reinforcement learning is basically about letting computers learn on their own by learning from themselves. Like it's said, you can only be as good as your master. Therefore if a computer learns from the world's best chess player then the best it can become is to become equal to the best chess player but to surpass her, AI needs to learn much more than just from the best chess player, and that is made possible by learning from itself, allowing it to take random decisions and then regarding the decisions which lead to a positive outcoming and punishing for decisions which led to a negative outcome, and rewarding the AI for not just winning the war but when it wins the battle too. This learning from itself is called reinforcement learning.
    04:07 : The difference between Reinforcement and SUpervised learning is that unlike in supervised learning where we need a training set like the moves of the best chess player to train our AI, and then the computer recognizes patterns picks the best pattern.
    In reinforcement learning, there is no training data and the computer pretty much learns by taking random decisions and figuring out which random move worked best
    04:41 : Policry gradients- AI does a random action >>checks if it is good>> if good asks it to repeat it and reward it>>if not, then punish it
    05:02 : 📌the entire goal of the policy network is to maximize the reward. It just receives the scoreboard as a checking mark
    05:37 : 📌read Andrej Karpath's blog on dep reinforcement learning: pong from pixels
    07:50 : the problem with policy gradient is that it rewards the end goal and not the process. so even if the AI took all the right steps in the game but only lost out on the last move, then the policy gradient would put all the moves made in the game as negative and will punish the ai for it. This problem is called the "Credit assignment Problem' To correct this, the AI can be rewarded for all the right moves in the game rather than winning the game.
    The solution given is called reward shaping.
    but the problem with reward shaping is that it has to be configured for all the cases where it is used. therefore makes it difficult to be used universally
    12:03 : Reward shaping can also have "the alignment problem" where the ai is getting all the rewards but isn't doing what it is supposed to do
    14:08 : Boston Dynamics has some pretty cool robots but those robots cant take autonomous intelligent decisions. They pre-programmed for doing what they do. They don't actively decide for themselves what they want to do. Hence they are not really intelligent and just a marketing gimmick at this point

  • @papaman1037
    @papaman1037 6 років тому +1

    Your content is far better than that guy that copies someone's code from GitHub makes an obscure reference to the original author and states that he added a wrapper to make the code easier to use (a lie Everytime I've checked). He uploads the code as an original comit (no fork from the rightful author's repo). He intentionally misleads people and profits from it -- a legal necessity for calling it fraudulent.
    Your content is excellent, clearly founded in recent research papers and you very professionally point out that material and more. You add value with your discussion of the topic. Thank you for an excellent channel. I would use patreon but I am Ill and not working. I'm doing my best to spread the word.

  • @williamkyburz
    @williamkyburz 6 років тому +1

    Xander, extremely well done, lucid and cogent. You should be teaching at M.I.T. or Universiteit Gent). The ability to teach complex subjects in an intuitive and simple way is a gift. Wish you the best in everything. Peace

    • @ArxivInsights
      @ArxivInsights  6 років тому +1

      Thanks William! I am actually doing my PhD in Gent at the moment :)

  • @MuditBachhawatIn
    @MuditBachhawatIn 4 роки тому

    I have been meaning to read about RL for a long time. This video couldn't be more simple and clear introduction to it. Thanks man!

  • @HARtalks
    @HARtalks 4 роки тому

    It was really interesting and helped me to get a clear picture of what reinforcement learning is... Thank you!!

  • @espangie
    @espangie 4 роки тому +1

    This was really helpful. Thank you to people like you for creating this content. Appreciate you, Xander!

  • @ArnauViaMartinezSeara
    @ArnauViaMartinezSeara 6 років тому

    Really useful. I am preparing a Reinforcement Learning class aplied to finance and it is really helpful. Can't wait to see next episode. Thanks

  • @gusbakker
    @gusbakker 5 років тому +2

    Great balance between a very well explained content and the interesting facts about current progress in AI at the end. Good work

  • @gudusangtani
    @gudusangtani 4 роки тому

    So well explained ....I also liked the comments on Boston robotics considering the hype and buzz about AI and ML.. You are doing a very good job !

  • @mujahid1324
    @mujahid1324 4 роки тому

    I would say "Wow'. You nailed it in10 mnts what's "reinforcement learning" is. Please keep sending more and more Ai . keep it up, Xander :)

  • @shashankshivakumar4732
    @shashankshivakumar4732 5 років тому +4

    I love this video. I love his criticial and grounded thinking. Great work !

  • @govindnarasimman1536
    @govindnarasimman1536 5 років тому

    Very clear naration and true to.ground comments. All the euphoria about AI needs to be grounded.

  • @majeedhussain3276
    @majeedhussain3276 6 років тому

    You deserve million subscribers hopefully one day you will. So much clarity in every video. Keep going...

  • @laeeqahmed1980
    @laeeqahmed1980 5 років тому +1

    Great talk. Humans are not good at multiple sound recognition and you added music to your video.

  • @josefpolasek6666
    @josefpolasek6666 4 роки тому +1

    Your videos are absolutely amazing! Thank you very much for explaining concept of RL in 16 minutes.

  • @Krimson5pride
    @Krimson5pride 5 років тому

    It was both professional and entertaining at the same time. Great and precise explanation.

  • @poojanpatel2437
    @poojanpatel2437 6 років тому +4

    Best Channel on yt for ml/dl/rl/ai... Keep up the good work... Would love to see your new video weekly...

    • @ArxivInsights
      @ArxivInsights  6 років тому +3

      I'd love to make more videos too! But since I'm currently doing this 100% in my spare time and 1 vid takes about 30hrs of work, there's really no way I can do one per week for now :(

    • @poojanpatel2437
      @poojanpatel2437 6 років тому

      Arxiv Insights Still amazing work till now... Love to see your more videos in future.. ❤

  • @nateshrager512
    @nateshrager512 6 років тому

    Great job introducing the topic. Very nice job dispelling misconceptions surrounding the topic as well. I put on that notification for your next videos, looking forward to em : )

  • @doctorartin
    @doctorartin 5 років тому

    Doing part of my PhD on potantial AI-strategies fordecision-making in healthcare, and this was very useful, thank you.

  • @thanasispappas62
    @thanasispappas62 Рік тому

    By far the best video of RL ive ever seen.

  • @Z4NT0
    @Z4NT0 3 роки тому

    I learned so much in just 16 minutes. Awesome Video!

  • @PasseScience
    @PasseScience 4 роки тому +2

    Hello, an RL idea I had, I am curious to know if you came accross similar things.
    Let's put the context in a very general way: a predictive/policy part doing it's usual job: having a latent/feature representation of the time line. (this time line including sense-data and action outputs, both past and predicted).
    and an RL part that can use the prediction of the policy part to make decisions. (determine action outputs).
    If we remain general both parts are working in some kind of a loop (policy predicts a future, decision parts tries to use it to determine futur actions, policy predicts again a futur based on what is planned etc...)
    We have here a very basic feature we usually seek: action that requires initially a huge number of back and forth in prediction-decision, can be eventually learned by the prediction parts (that will prefill the output
    actions based on it's prediction). Here nothing new, I am just talking about an abstraction that can suit a large number of RL systems.
    But now is the idea: usually the decision part job is to fill action output, but if we allow it to fill the "sense data prediction part" we end up with something interesting: we can see the prediction part
    in a more general way, not only as something that can predict but as something that can fill the gap (prediction is filling a specific gap), and so if the decision parts prefil the predicted sense-data with "i fill an apple in my hand" the predictive part (now more a "time-line filling gap part") can try to determine the actions that leads to this sense-data. Here we invent a new way of decision to communicate: by "will". It describes what it wants and the planification to get it is delegated to the first part of the engine.
    Was I clear?
    have you seen this kind of thing?

  • @LearnRoboticsAndAI
    @LearnRoboticsAndAI 3 роки тому

    Summary :
    - State-of-the-art Robotics is a Software challenge and not a hardware challenge (Robots are physically capable of challenging tasks)
    - Supervised Learning
    * Known - Inputs, Outputs
    * Compute gradients using Backpropagation to train the network to predict outputs for new inputs
    * E.g. for a game of ping pong, the data can be screenshots at specific time instants and the key (Up/down) pressed at each instant by the user (recorded from a user playing the game) and it can be used to train a neural network to predict output for a new input image
    * Disadvantage is the creation of a dataset, which isn't always easy to do
    * Another disadvantage is that since the data is recorded from human playing the game, it can never be better than a human playing the game
    - Reinforcement Learning
    * Difference to supervised learning is that we do not know the target label as we have no dataset
    * The network that transforms the input states to output actions is called policy network
    * One simple way to train a policy network (can be fully connected or convoluted) is a method called policy gradients
    1. In Policy Gradient method, we start with a completely random network
    2. Feed the network a frame from the game engine, it produces a random output
    3. Send that action back to the game engine
    4. The game engine produces the next frame
    5. Outputs are represented probabilistically sampled from a distribution such that the same exact actions are not repeated again and again
    6. Rewards are given if the agent scores a goal (+1) and penalty is given is the opponent scores a goal (-1)
    7. Entire goal is to optimize its policy to receive maximum reward
    8. We get a bunch of experience by feeding frames to the network, getting random actions
    9. Sometimes (rarely) the result would be a WIN
    10. We use normal gradients to increase the probability of those actions (that result in WIN) in the future
    11. For a negative reward we use the same gradient but multiply it with -1
    * Credit Assignment Problem - Most of the steps were good but it lost at the end so our network will think that the particular sequence of actions is bad
    * Sparse reward setting, very sample inefficient
    * Reward shaping- Additional intermediate rewards. But must be designed individually for specific problems. Not scalable.

  • @7810
    @7810 6 років тому

    Good stuff to learn the RL in terms of basic knowledge as well as the challenge it will face. Thanks for your time and sharing!

  • @HarutakaShimizu
    @HarutakaShimizu 6 місяців тому

    Wow, this was a very clearly explained video, thanks!

  • @ms_1918
    @ms_1918 5 років тому

    well came here for a 1 min intro to reinforcement learning for first class of course,
    stopped after 16 minutes what a superb experience.

  • @rishidixit7939
    @rishidixit7939 Рік тому +2

    The sudden surprise of hearing Bruno Mars makes you pause video for other open tabs

  • @alirezaparsay8518
    @alirezaparsay8518 Рік тому

    The explanation was so clear. Thank you.

  • @dean8147
    @dean8147 3 роки тому

    You’re a legend mate. Honestly, thanks for all of your hard work

  • @jackwhite9332
    @jackwhite9332 6 років тому +7

    Impressive explanation, found this very useful. Thank you!

  • @nemx4u
    @nemx4u 6 років тому +2

    You explain hard topics beautifully! great job. Would love to see more RL videos!

  • @orfeasliossatos
    @orfeasliossatos 6 років тому +2

    I've been literally looking all over for a video like this, thank you so much

  • @shirishbajpai9486
    @shirishbajpai9486 Рік тому

    watched in 2023 after all the LLMs stuff going on... still such relevant and pure gold!

  • @amitredkar140
    @amitredkar140 6 років тому +1

    Great video!!!! Explained exceptionally, liked other videos as well from your channel. Would love to see more stuff related to AI/DL or RL. Thanks in advance. Keep up the good work....

  • @ArturoMoraSoto
    @ArturoMoraSoto 4 роки тому

    Nice explanation, thanks for taking the time to create this great video.

  • @steadymedia234
    @steadymedia234 5 років тому

    This is a great presentation on RL, short and clear content.

  • @codyheiner3636
    @codyheiner3636 6 років тому +1

    Love the philosophical discussion at the end!

  • @DotCSV
    @DotCSV 6 років тому +46

    Hi Xander, just found your UA-cam channel and I'm very amazed about your content! I also run a UA-cam channel with the same topic but for the Spanish speaking audience, and I'm happy to see that more new channels are growing to educate in the field of machine learning. I hope in the future we can crossover our contents :)

    • @ArxivInsights
      @ArxivInsights  6 років тому +11

      Checked out your channel, great stuff man!! It's indeed nice to see that many people are starting to contribute to the online ML community in such a huge variety of ways :p

    • @OcramRatte
      @OcramRatte 4 роки тому

      eeee yo creo que te acabo de ver en tiktok

  • @sharadrawatindia
    @sharadrawatindia 6 років тому

    Hey Xander! Great videos. Looking forwards for your next video.

  • @soumyakantadash5986
    @soumyakantadash5986 5 років тому

    These videos are gem!!!..... incredible, precise and knowledgeable!!!!

  • @RoxanaNoe
    @RoxanaNoe 6 років тому

    Your channel is a great resource for getting into Deep Learning and AI.

  • @ahilanpalarajah3159
    @ahilanpalarajah3159 5 років тому

    Only way to describe this guy is "22 Two's - Jay-Z". Excellent video.

  • @bsudharsh
    @bsudharsh 5 років тому

    succinct; its a brilliant rendition on reinforcement learning

  • @mantische
    @mantische 4 роки тому

    One of the best explanations I've seen

  • @mohammadhatoum
    @mohammadhatoum 6 років тому

    Great job.. Explained the subject in a simple way. Keep it up and looking forward for new videos

  • @mehdisauvage1234
    @mehdisauvage1234 6 років тому

    Your videos are so useful and interesting ! This is pure gold to me :)

  • @Jshizzle2
    @Jshizzle2 5 років тому +1

    Perfect video, so much more intuitive than my lectures. Thanks a bunch!

  • @sidharthaparhi7930
    @sidharthaparhi7930 6 років тому

    Also your intro is very high quality, like an intro to a good TV show

  • @OliverZeigermann
    @OliverZeigermann 6 років тому

    Very lively and understandable. Great work!

  • @khajasaen
    @khajasaen 6 років тому

    Best channel in the crowd ... keep it up Xander

  • @geraldkenneth119
    @geraldkenneth119 2 роки тому +1

    It seems to me one way, albeit a rather difficult one, to help AI deal with sparse rewards is to
    1. Give them a reward function that doesn’t work based on if they accomplished the task or not, but on how close they got to achieving it
    2. Give them the ability to generate plans for achieving a goal, and to recognize why they failed

  • @josephedappully1482
    @josephedappully1482 6 років тому

    This is a great video; thanks for making it! Looking forward to your next one.

  • @ingeniouswild
    @ingeniouswild 6 років тому +1

    Very nice episode! One thing that struck me about your suggestion that without Reward Shaping, the auto-learning of the 2600 games would be intractable: even for a human, this would be extremely difficult - we succeed with new, undocumented games because they often have similar sub-components and sub-goals that we already know from other games (or life). But I'm sure you could easily construct a game which would be impossible for a human to learn without any hints, while still having the same overall complexity.

  • @saaniausaf9621
    @saaniausaf9621 6 років тому

    I loved the way you explained everything. Thanks!

  • @alenasazanova8331
    @alenasazanova8331 4 роки тому

    That's very interesting and understantable video. Thank you very much!

  • @tnmygrwl
    @tnmygrwl 6 років тому

    You do an awesome of structuring the content. Loved the video.

  • @wzyjoseph7317
    @wzyjoseph7317 3 роки тому

    Very clear explaination! Thanks for the work!!!!XD

  • @papaman1037
    @papaman1037 6 років тому

    Even in games w/o an RL actor loops without achieving a goal occur. The long time solution was to periodically perturb the system sufficiently that such learned patterns get interrupted.

  • @alanator25
    @alanator25 Рік тому

    Thank you! This was a great introduction!

  • @sridhasridharan3600
    @sridhasridharan3600 4 роки тому

    Great Videos! I am recommending these to my students.

  • @Alex-gc2vo
    @Alex-gc2vo 6 років тому

    your videos are some of the best explanations I've found for a lot of these very advanced subjects. I suspect your viewer count is going to jump very quickly. keep it up.

  • @DavidSaintloth
    @DavidSaintloth 6 років тому

    Reinforcement learning is along the path to the complex multidimensional salience models that will drive dynamic cognition.
    "Reward shaping" I assert in the salience theory of dynamic cognition that I proposed publicly in 2013 is performed by a combination of autonomic and emotional signal modifications to experience. The key is to tie the reward to the experience and then use that to vary the prediction...this way you don't reward shape as a separate process ...reward shaping is actually performed BY comparison.
    For those interested in the salience theory of dynamic cognition and consciousness a collection of the articles I've written are available at this public Facebook note:
    facebook.com/notes/david-saintloth/discovering-the-dynamic-cognition-cycle/10152513149708057

  • @azmathmoosa4324
    @azmathmoosa4324 6 років тому

    I like how u don't hype up anything. Great mate! I subscribe!

  • @colorlace
    @colorlace 5 років тому +16

    The Lebowski Theorem: No superintelligent AI is going to bother with a task that is harder than hacking its reward function.

    • @wizardOfRobots
      @wizardOfRobots 4 роки тому +6

      Unless it's reward function punishes it for it.
      Now we have the Meta-Lebowski theorem: It's not going to bother with a task harder than hacking it's hack-detection algorithm.

    • @halifakx
      @halifakx 3 роки тому

      perhaps, a machine become smart, and then smarter as it decides becoiming smarter is shorterst path to reward... finally so smart to realize their reward is just color mirrors? and create a new program inside the program that cancels or outweigh the previous reward and create new rewards? programming this new reawards in their own languaje, not apparent to us....like facebook robot talking their own languaje

    • @halifakx
      @halifakx 3 роки тому

      estramboticusssssss dangerosicusss hahaha

  • @robertfairburn9979
    @robertfairburn9979 6 років тому

    When I was a psychology student when trained chickens using reinforcement training with reward shaping. However it was a form supervised training in reality

  • @thaermashkoor6225
    @thaermashkoor6225 3 роки тому

    Thanks for this clear introduction.

  • @papaguagua
    @papaguagua 2 роки тому

    I love the use of Xi when you talk about monopolies. Lol

  • @luiseduardocorralesmendoza9396
    @luiseduardocorralesmendoza9396 5 років тому

    Great examples and great explanation, thank you i was struggling with this topic

  • @bjbodner3097
    @bjbodner3097 6 років тому

    Great video, great channel!
    Thanks so much for making this!
    Can't wait to watch more:)

  • @empiricistsacademy7181
    @empiricistsacademy7181 6 років тому

    Thanks youuu for this video. Looking forward to your future videos!

  • @jorgegarcia-torresfdez2471
    @jorgegarcia-torresfdez2471 6 років тому

    You did again a really nice work ! Congratulations :D

  • @maisamwasti
    @maisamwasti 6 років тому

    Your videos = super informative! Thanks a lot for the good work

  • @jonathaskerber5472
    @jonathaskerber5472 5 років тому

    Such a great introduction. Keep up the good work!

  • @tonakkie635
    @tonakkie635 5 років тому +1

    Great overview, well explained👍.Thanks

  • @rajendrarao3057
    @rajendrarao3057 6 років тому

    awesome video sir. please keep up the good work in this field....

  • @stefano3808
    @stefano3808 4 роки тому

    really high quality videos, thanks for that

  • @arjunpatel3699
    @arjunpatel3699 4 роки тому

    Subscribed! I am looking forward to your videos.

  • @digvijaybhandari9747
    @digvijaybhandari9747 Рік тому

    Really enjoyed the content here!

  • @JuantheJohn1
    @JuantheJohn1 2 роки тому

    damn 10:50 went from "damn I'm in class" to "YOO IM AT THE PARTY"