Making Real-World Reinforcement Learning Practical

Поділитися
Вставка
  • Опубліковано 2 січ 2024
  • Lecture by Sergey Levine about progress on real-world deep RL. Covers these papers:
    A Walk in the Park: Learning to Walk in 20 Minutes With Model-Free Reinforcement Learning: sites.google.com/berkeley.edu...
    Grow Your Limits: Continuous Improvement with Real-World RL for Robotic Locomotion: sites.google.com/berkeley.edu...
    Reset-Free Reinforcement Learning via Multi-Task Learning: Learning Dexterous Manipulation Behaviors without Human Intervention: sites.google.com/view/mtrf
    REBOOT: Reuse Data for Bootstrapping Efficient Real-World Dexterous Manipulation: sites.google.com/view/reboot-...
    FastRLAP: A System for Learning High-Speed Driving via Deep RL and Autonomous Practicing: sites.google.com/view/fastrlap
    Q-Transformer: Scalable Offline Reinforcement Learning via Autoregressive Q-Functions: qtransformer.github.io/
    Deep RL at Scale: Sorting Waste in Office Buildings with a Fleet of Mobile Manipulators: rl-at-scale.github.io/
  • Розваги

КОМЕНТАРІ • 11

  • @user-to9ub5xv7o
    @user-to9ub5xv7o 4 місяці тому +3

    Chapter 1: Introduction to Reinforcement Learning in the Real World [0:00-0:54]
    - Sergey Lan discusses the importance of reinforcement learning (RL) in AI, differentiating it from generative AI techniques.
    - Highlights the ability of RL to achieve results beyond human capabilities, using the example of AlphaGo's unique moves.
    Chapter 2: Advantages of Real World RL and Challenges in Simulation [0:54-2:13]
    - Emphasizes the power of RL in real-world scenarios over simulation.
    - Discusses the challenges in building accurate simulators, especially for complex environments involving human interaction.
    Chapter 3: Reinforcement Learning in Complex Environments [2:13-3:18]
    - Describes how RL is more effective in complex, variable environments.
    - Explains the need for advanced learning algorithms to interact with these environments for emerging behaviors.
    - Argues for RL's potential in optimizing policies specifically for real-world settings.
    Chapter 4: Progress and Practicality of Real-World Deep RL [3:18-5:52]
    - Outlines the advancements in deep RL, making it more practical and scalable in the real world.
    - Details on how sample complexity has significantly improved, allowing for real-world locomotion skills to be learned in minutes.
    - Discusses leveraging prior data in RL and overcoming challenges previously considered as showstoppers.
    Chapter 5: Learning Locomotion Skills in Real World with Deep RL [5:52-11:46]
    - Sergey shares examples of learning locomotion skills in real robots, progressing from simplified models to more complex ones.
    - Highlights the significant reduction in training time due to improved algorithms and engineering techniques.
    - Discusses the importance of regularization and taking numerous gradient steps for fast learning.
    Chapter 6: Manipulation Skills Through Real World Experience [11:46-16:03]
    - Transition to learning manipulation skills in real-world settings.
    - Discusses the challenges and solutions for continuous learning without human intervention.
    - Explains the integration of multiple tasks to enable autonomous learning and resetting.
    Chapter 7: Leveraging Heterogeneous Prior Data in RL [16:03-21:40]
    - Focuses on combining data-driven learning with RL to improve efficiency.
    - Describes how pre-training with diverse data can lead to rapid adaptation in new tasks.
    - Uses examples to illustrate the efficiency of RL with prior data in both locomotion and manipulation tasks.
    Chapter 8: Scaling Deep RL for Practical Applications [21:40-26:30]
    - Sergey discusses the scalability of deep RL in practical, real-world applications.
    - Provides examples of real-world deployments, including navigation and waste sorting robots.
    - Emphasizes the importance of continuous improvement and adaptation in diverse environments.
    Chapter 9: Future Directions and Potential of Real World RL [26:30-38:20]
    - Concludes with a discussion on future potentials and improvements needed in real world RL.
    - Suggests areas for further research, including leveraging prior data for exploration and lifelong learning.
    - Acknowledges the need for more stable, efficient, and reliable RL methods for mainstream adoption.

  • @chillmathematician3303
    @chillmathematician3303 4 місяці тому +7

    We need to make people focus back to RL again

  • @sapienspace8814
    @sapienspace8814 4 місяці тому +2

    Very interesting talk, thank you for sharing!
    I used to work in Reinforcement Learning (RL) over 26 years ago on an inverted pendulum control system that was first simulated, then applied to a real pendulum. I was amazed at how well and how quickly it was able to learn how to balance. It seemed a medium resolution of the state space was best for learning rates. I had Dr. Barto's and Dr. Sutton's book, the first edition, on Reinforcement Learning that was released in 1997.
    Did some k-means clustering on the input space to focus attention of learning on regions of interest in the state space and also used Fuzzy Logic (a sort of merging with language and math to Fuzzify and De-Fuzzify the state space) with Radial Basis Functions as approximators.
    With the seemingly sudden take off of RL again now in the last few years, I have now bought their 2nd edition of RL and am slowly going through it, also bought their key reference "The Hedonistic Neuron" by Klopf.

  • @godwyllaikins3277
    @godwyllaikins3277 4 місяці тому +2

    This was truly an amazing presentation.

  • @NoNTr1v1aL
    @NoNTr1v1aL 4 місяці тому +3

    Absolutely brilliant talk!

  • @lemurpotatoes7988
    @lemurpotatoes7988 Місяць тому

    Extremely interesting, the part about being less aggressive when prediction is harder kind of reninds me of some speculative thoughts I had about artificial neurotransmitters aping the locus coruleus norepinephrine system to control error rates of subcircuits to speed up learning, that idea is still underdeveloped though and I was thinking more about optimizing the joint relationships of different routes into a neuron which would have atrocious scaling done naively.

  • @ArtOfTheProblem
    @ArtOfTheProblem 4 місяці тому +1

    thank you for sharing

  • @yurakriachko7008
    @yurakriachko7008 4 дні тому

    A very nice overview. What is a good place to learn it from code examples?

  • @zabean
    @zabean 4 місяці тому

    thank you mate

  • @rfernand2
    @rfernand2 4 місяці тому +4

    Thanks - I haven't been tracking RL for past several years, so this is a nice high-level update on things, with linked papers for details. Given this progress, are we about to see an explosion in robotics deployment? If so, will it be mostly industrial, or some consumer impact also?

    • @Alex-kr7zr
      @Alex-kr7zr 3 дні тому

      I'd bet that the consumer impact will be there before this will be picked up by "industrial" applications. Industrial applications are so stuck in the 70s style static robot programming with SPS and ladder logic that they will be extremely late to the game compared to startups and research already working with AI/ML. Also, this will have the most impact on small to medium size companies. Big companies can already afford an expensive robot programmer to the do the programming of the industrial robot, small companies simply can't, so they will benefit most from easier (and therefore much cheaper) robot programming. The hardware might still be expensive, but the current reality is that programming a robot is much more expensive than the hardware (even something like a UR isn't fully end user friendly and requires expert knowledge of some level).